@@ -140,7 +140,7 @@ TEST_CASE_METHOD(simple_connected_test, "inserting values using the simple api w
140
140
db.insert (line (" test_timestamp" , key_value_pairs (" a1" , " b2" ), key_value_pairs (" c3" , " d4" ), dummy));
141
141
dummy.stamp = " 97553045000000000" ; // 1973-02-03
142
142
db.insert (line (" test_timestamp" , key_value_pairs (" e1" , " f2" ), key_value_pairs (" g3" , " h4" ), dummy));
143
-
143
+
144
144
wait_for ([] {return false ; }, 3 );
145
145
146
146
auto res = result (" test_timestamp" );
@@ -157,10 +157,10 @@ SCENARIO_METHOD(simple_connected_test, "more than 1000 inserts per second") {
157
157
GIVEN (" A connection to the db" ) {
158
158
influxdb::async_api::simple_db asyncdb (" http://localhost:8086" , db_name);
159
159
using Clock = std::chrono::high_resolution_clock;
160
- auto many_times = 123456_times ;
160
+ auto many_times = 111000_times ;
161
161
162
162
// https://www.influxdata.com/influxdb-1-1-released-with-up-to-60-performance-increase-and-new-query-functionality/
163
- const int MAX_VALUES_PER_TAG = 100000 ;
163
+ const int MAX_VALUES_PER_TAG = 50000 ; // actually, 100000
164
164
165
165
WHEN (" I send a large number of unique entries" ) {
166
166
auto t1 = Clock::now ();
@@ -183,7 +183,7 @@ SCENARIO_METHOD(simple_connected_test, "more than 1000 inserts per second") {
183
183
AND_THEN (" All entries arrive at the database" ) {
184
184
// wait for asynchronous fill
185
185
auto query = std::string (" select count(*) from " ) + db_name + " ..asynctest" ;
186
- wait_for ([this , query, many_times] { return raw_db.get (query).find (std::to_string (many_times.count )) != std::string::npos; }, 100 );
186
+ wait_for ([this , query, many_times] { return raw_db.get (query).find (std::to_string (many_times.count )) != std::string::npos; }, 200 );
187
187
bool all_entries_arrived = raw_db.get (query).find (std::to_string (many_times.count )) != std::string::npos;
188
188
189
189
CHECK (all_entries_arrived);
0 commit comments