You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
5718 lines
595 KiB
5718 lines
595 KiB
18:07:49,354 graphrag.index.cli INFO Logging enabled at E:\uc\Source-Web-23070-DataAnalysis\python\output\indexing-engine.log
|
|
18:07:49,358 graphrag.index.cli INFO Starting pipeline run for: 20250623-180749, dryrun=False
|
|
18:07:49,359 graphrag.index.cli INFO Using default configuration: {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"root_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python",
|
|
"reporting": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"storage": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"cache": {
|
|
"type": "file",
|
|
"base_dir": "cache",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"input": {
|
|
"type": "file",
|
|
"file_type": "text",
|
|
"base_dir": "input",
|
|
"storage_account_blob_url": null,
|
|
"encoding": "utf-8",
|
|
"file_pattern": ".*\\.txt$",
|
|
"file_filter": null,
|
|
"source_column": null,
|
|
"timestamp_column": null,
|
|
"timestamp_format": null,
|
|
"text_column": "text",
|
|
"title_column": null,
|
|
"document_attribute_columns": []
|
|
},
|
|
"embed_graph": {
|
|
"enabled": false,
|
|
"num_walks": 10,
|
|
"walk_length": 40,
|
|
"window_size": 2,
|
|
"iterations": 3,
|
|
"random_seed": 597832,
|
|
"strategy": null
|
|
},
|
|
"embeddings": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_embedding",
|
|
"model": "text-embedding-3-small",
|
|
"max_tokens": 4000,
|
|
"temperature": 0,
|
|
"top_p": 1,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": null,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"batch_size": 16,
|
|
"batch_max_tokens": 8191,
|
|
"target": "required",
|
|
"skip": [],
|
|
"vector_store": null,
|
|
"strategy": null
|
|
},
|
|
"chunks": {
|
|
"size": 1200,
|
|
"overlap": 100,
|
|
"group_by_columns": [
|
|
"id"
|
|
],
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"snapshots": {
|
|
"graphml": false,
|
|
"raw_entities": false,
|
|
"top_level_nodes": false
|
|
},
|
|
"entity_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/entity_extraction.txt",
|
|
"entity_types": [
|
|
"organization",
|
|
"person",
|
|
"geo",
|
|
"event"
|
|
],
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"summarize_descriptions": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/summarize_descriptions.txt",
|
|
"max_length": 500,
|
|
"strategy": null
|
|
},
|
|
"community_reports": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/community_report.txt",
|
|
"max_length": 2000,
|
|
"max_input_length": 8000,
|
|
"strategy": null
|
|
},
|
|
"claim_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"enabled": false,
|
|
"prompt": "prompts/claim_extraction.txt",
|
|
"description": "Any claims or facts that could be relevant to information discovery.",
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"cluster_graph": {
|
|
"max_cluster_size": 10,
|
|
"strategy": null
|
|
},
|
|
"umap": {
|
|
"enabled": false
|
|
},
|
|
"local_search": {
|
|
"text_unit_prop": 0.5,
|
|
"community_prop": 0.1,
|
|
"conversation_history_max_turns": 5,
|
|
"top_k_entities": 10,
|
|
"top_k_relationships": 10,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"llm_max_tokens": 2000
|
|
},
|
|
"global_search": {
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"data_max_tokens": 12000,
|
|
"map_max_tokens": 1000,
|
|
"reduce_max_tokens": 2000,
|
|
"concurrency": 32
|
|
},
|
|
"encoding_model": "cl100k_base",
|
|
"skip_workflows": []
|
|
}
|
|
18:07:49,362 graphrag.index.create_pipeline_config INFO skipping workflows
|
|
18:07:49,362 graphrag.index.run.run INFO Running pipeline
|
|
18:07:49,362 graphrag.index.storage.file_pipeline_storage INFO Creating file storage at E:\uc\Source-Web-23070-DataAnalysis\python\output
|
|
18:07:49,363 graphrag.index.input.load_input INFO loading input from root_dir=input
|
|
18:07:49,363 graphrag.index.input.load_input INFO using file storage for input
|
|
18:07:49,365 graphrag.index.storage.file_pipeline_storage INFO search E:\uc\Source-Web-23070-DataAnalysis\python\input for files matching .*\.txt$
|
|
18:07:49,367 graphrag.index.input.text INFO found text files from input, found [('10_7\u4f5c\u696d\u5206\u4eab.txt', {}), ('\u4f5c\u54c1\u5206\u4eab(09_30).txt', {}), ('\u4f5c\u54c1\u5206\u4eab\uff08\u9ede\u6211\u9032\u5165\u7de8\u8f2f\uff09.txt', {}), ('\u7b2c\u4e00\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1113-14).txt', {}), ('\u7b2c\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e03\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e09\u6b21office hour(1128-1129).txt', {}), ('\u7b2c\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e5d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e8c\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1120-1122).txt', {}), ('\u7b2c\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516b\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {})]
|
|
18:07:49,390 graphrag.index.input.text INFO Found 22 files, loading 22
|
|
18:07:49,394 graphrag.index.workflows.load INFO Workflow Run Order: ['create_base_text_units', 'create_base_extracted_entities', 'create_summarized_entities', 'create_base_entity_graph', 'create_final_entities', 'create_final_nodes', 'create_final_communities', 'create_final_relationships', 'create_final_text_units', 'create_final_community_reports', 'create_base_documents', 'create_final_documents']
|
|
18:07:49,394 graphrag.index.run.run INFO Final # of rows loaded: 22
|
|
18:07:49,550 graphrag.index.run.workflow INFO dependencies for create_base_text_units: []
|
|
18:07:49,559 datashaper.workflow.workflow INFO executing verb orderby
|
|
18:07:49,569 datashaper.workflow.workflow INFO executing verb zip
|
|
18:07:49,577 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
18:07:49,590 datashaper.workflow.workflow INFO executing verb chunk
|
|
18:07:51,978 datashaper.workflow.workflow INFO executing verb select
|
|
18:07:51,984 datashaper.workflow.workflow INFO executing verb unroll
|
|
18:07:51,993 datashaper.workflow.workflow INFO executing verb rename
|
|
18:07:51,998 datashaper.workflow.workflow INFO executing verb genid
|
|
18:07:52,8 datashaper.workflow.workflow INFO executing verb unzip
|
|
18:07:52,15 datashaper.workflow.workflow INFO executing verb copy
|
|
18:07:52,21 datashaper.workflow.workflow INFO executing verb filter
|
|
18:07:52,36 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_text_units.parquet
|
|
18:07:52,225 graphrag.index.run.workflow INFO dependencies for create_base_extracted_entities: ['create_base_text_units']
|
|
18:07:52,226 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:07:52,272 datashaper.workflow.workflow INFO executing verb entity_extract
|
|
18:07:52,288 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:07:52,298 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for gpt-4-turbo-preview: TPM=0, RPM=0
|
|
18:07:52,298 graphrag.index.llm.load_llm INFO create concurrency limiter for gpt-4-turbo-preview: 25
|
|
18:07:54,337 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:07:54,341 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.9380000000819564. input_tokens=2188, output_tokens=5
|
|
18:07:55,911 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:07:55,912 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.5309999994933605. input_tokens=2937, output_tokens=65
|
|
18:07:57,233 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:07:57,235 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.843999999575317. input_tokens=2938, output_tokens=90
|
|
18:07:58,36 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:07:58,38 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.8119999999180436. input_tokens=1806, output_tokens=5
|
|
18:07:58,351 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:07:58,354 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.890999999828637. input_tokens=1997, output_tokens=109
|
|
18:08:00,82 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:00,83 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.734000000171363. input_tokens=2936, output_tokens=192
|
|
18:08:00,540 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:00,543 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.061999999918044. input_tokens=2936, output_tokens=152
|
|
18:08:01,825 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:01,827 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.375. input_tokens=2936, output_tokens=219
|
|
18:08:02,141 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:02,142 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.688000000081956. input_tokens=2937, output_tokens=228
|
|
18:08:02,159 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:02,161 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.796000000089407. input_tokens=2936, output_tokens=148
|
|
18:08:04,149 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:04,151 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.70299999974668. input_tokens=2936, output_tokens=262
|
|
18:08:04,247 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:04,249 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.764999999664724. input_tokens=2936, output_tokens=226
|
|
18:08:04,939 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:04,940 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.57799999974668. input_tokens=2936, output_tokens=232
|
|
18:08:05,49 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:05,50 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.67200000025332. input_tokens=2936, output_tokens=169
|
|
18:08:05,314 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:05,316 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.1569999996572733. input_tokens=1970, output_tokens=5
|
|
18:08:05,666 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:05,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.57799999974668. input_tokens=2936, output_tokens=125
|
|
18:08:06,975 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:06,976 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.828999999910593. input_tokens=2937, output_tokens=120
|
|
18:08:07,58 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:07,60 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.625. input_tokens=2936, output_tokens=198
|
|
18:08:07,569 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:07,572 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.234000000171363. input_tokens=2937, output_tokens=329
|
|
18:08:07,579 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:07,580 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.203000000678003. input_tokens=2936, output_tokens=211
|
|
18:08:08,667 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:08,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.093999999575317. input_tokens=2097, output_tokens=5
|
|
18:08:08,790 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:08,794 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.436999999918044. input_tokens=2936, output_tokens=243
|
|
18:08:09,171 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:09,173 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.75. input_tokens=2935, output_tokens=394
|
|
18:08:09,907 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:09,910 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.1090000001713634. input_tokens=2936, output_tokens=5
|
|
18:08:10,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:10,296 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.15599999949336. input_tokens=2937, output_tokens=138
|
|
18:08:13,604 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:13,605 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.938000000081956. input_tokens=2936, output_tokens=117
|
|
18:08:13,947 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:13,949 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.781000000424683. input_tokens=2937, output_tokens=122
|
|
18:08:14,327 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:14,328 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.0. input_tokens=2936, output_tokens=365
|
|
18:08:14,437 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:14,438 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.031999999657273. input_tokens=2936, output_tokens=366
|
|
18:08:14,450 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:14,452 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 6.875. input_tokens=2936, output_tokens=140
|
|
18:08:15,293 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:15,294 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.92200000025332. input_tokens=2936, output_tokens=388
|
|
18:08:15,618 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:15,619 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.375. input_tokens=2936, output_tokens=247
|
|
18:08:15,725 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:15,755 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.32799999974668. input_tokens=2935, output_tokens=384
|
|
18:08:16,471 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:16,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.42200000025332. input_tokens=2936, output_tokens=370
|
|
18:08:16,837 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:16,839 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.015999999828637. input_tokens=2937, output_tokens=262
|
|
18:08:17,374 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:17,375 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.014999999664724. input_tokens=2937, output_tokens=400
|
|
18:08:18,101 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:18,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.188000000081956. input_tokens=2904, output_tokens=558
|
|
18:08:18,664 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:18,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.20299999974668. input_tokens=2936, output_tokens=540
|
|
18:08:18,715 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:18,717 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.281000000424683. input_tokens=34, output_tokens=92
|
|
18:08:20,401 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:20,403 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.07799999974668. input_tokens=34, output_tokens=116
|
|
18:08:21,569 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:21,571 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.953000000678003. input_tokens=34, output_tokens=110
|
|
18:08:22,26 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:22,29 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.186999999918044. input_tokens=34, output_tokens=97
|
|
18:08:22,611 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:22,786 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.109000000171363. input_tokens=34, output_tokens=85
|
|
18:08:23,546 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:23,548 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.25. input_tokens=34, output_tokens=176
|
|
18:08:23,757 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:23,759 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.406000000424683. input_tokens=2937, output_tokens=488
|
|
18:08:25,876 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:25,877 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.764999999664724. input_tokens=34, output_tokens=213
|
|
18:08:26,453 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:26,461 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.015999999828637. input_tokens=34, output_tokens=246
|
|
18:08:26,818 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:26,820 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.218000000342727. input_tokens=34, output_tokens=278
|
|
18:08:26,936 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:26,938 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.890999999828637. input_tokens=2936, output_tokens=488
|
|
18:08:26,991 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:26,992 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.093999999575317. input_tokens=2938, output_tokens=373
|
|
18:08:27,957 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:27,959 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.42200000025332. input_tokens=34, output_tokens=103
|
|
18:08:28,172 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:28,178 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.125. input_tokens=2937, output_tokens=565
|
|
18:08:28,640 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:28,641 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.09400000050664. input_tokens=2935, output_tokens=635
|
|
18:08:28,910 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:28,912 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.436999999918044. input_tokens=34, output_tokens=171
|
|
18:08:29,979 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:29,987 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.688000000081956. input_tokens=2423, output_tokens=382
|
|
18:08:30,194 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:30,196 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.890000000596046. input_tokens=2936, output_tokens=537
|
|
18:08:30,676 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:30,678 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.70299999974668. input_tokens=2936, output_tokens=547
|
|
18:08:31,148 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:31,149 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.125. input_tokens=34, output_tokens=135
|
|
18:08:31,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:31,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 38.85900000017136. input_tokens=2937, output_tokens=754
|
|
18:08:31,460 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:31,465 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.890999999828637. input_tokens=34, output_tokens=146
|
|
18:08:31,871 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:31,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 39.40599999949336. input_tokens=2938, output_tokens=774
|
|
18:08:32,110 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:32,114 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.343999999575317. input_tokens=34, output_tokens=279
|
|
18:08:32,202 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:32,203 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.328000000678003. input_tokens=34, output_tokens=128
|
|
18:08:32,414 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:32,419 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.2659999998286366. input_tokens=34, output_tokens=5
|
|
18:08:33,899 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:33,902 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.07799999974668. input_tokens=34, output_tokens=192
|
|
18:08:34,40 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:34,42 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.67200000025332. input_tokens=34, output_tokens=291
|
|
18:08:34,96 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:34,98 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 29.156000000424683. input_tokens=2937, output_tokens=817
|
|
18:08:34,370 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:34,373 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.968999999575317. input_tokens=34, output_tokens=251
|
|
18:08:34,499 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:34,502 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.561999999918044. input_tokens=34, output_tokens=163
|
|
18:08:35,811 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:35,815 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 30.140999999828637. input_tokens=2936, output_tokens=729
|
|
18:08:36,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:36,801 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.609000000171363. input_tokens=34, output_tokens=123
|
|
18:08:36,936 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:36,938 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.063000000081956. input_tokens=34, output_tokens=140
|
|
18:08:36,964 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:36,965 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.04700000025332. input_tokens=34, output_tokens=118
|
|
18:08:37,816 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:37,818 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 19.109000000171363. input_tokens=34, output_tokens=375
|
|
18:08:38,248 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:38,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.4369999999180436. input_tokens=34, output_tokens=50
|
|
18:08:38,354 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:38,356 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.890999999828637. input_tokens=34, output_tokens=94
|
|
18:08:39,464 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:39,466 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.1090000001713634. input_tokens=1781, output_tokens=5
|
|
18:08:40,198 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:40,201 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 26.25. input_tokens=34, output_tokens=622
|
|
18:08:40,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:40,471 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.34400000050664. input_tokens=34, output_tokens=220
|
|
18:08:40,580 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:40,588 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.625. input_tokens=34, output_tokens=282
|
|
18:08:41,137 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:41,139 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.938000000081956. input_tokens=34, output_tokens=308
|
|
18:08:41,162 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:41,166 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.968999999575317. input_tokens=34, output_tokens=232
|
|
18:08:41,228 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:41,230 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.765999999828637. input_tokens=34, output_tokens=208
|
|
18:08:41,528 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:41,529 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.281000000424683. input_tokens=34, output_tokens=75
|
|
18:08:41,999 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:42,1 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.23399999924004. input_tokens=34, output_tokens=436
|
|
18:08:42,19 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:42,22 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.031000000424683. input_tokens=34, output_tokens=323
|
|
18:08:43,335 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:43,337 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.04700000025332. input_tokens=34, output_tokens=167
|
|
18:08:44,164 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:44,167 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.7029999997466803. input_tokens=2936, output_tokens=98
|
|
18:08:44,217 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:44,219 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.79700000025332. input_tokens=2936, output_tokens=258
|
|
18:08:44,624 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:44,626 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 21.82799999974668. input_tokens=34, output_tokens=514
|
|
18:08:45,926 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:45,927 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.296999999321997. input_tokens=34, output_tokens=283
|
|
18:08:46,18 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:46,22 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.561999999918044. input_tokens=34, output_tokens=117
|
|
18:08:46,288 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:46,289 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.125. input_tokens=34, output_tokens=80
|
|
18:08:46,560 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:46,581 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.890000000596046. input_tokens=34, output_tokens=384
|
|
18:08:46,595 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:46,597 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.21900000050664. input_tokens=2938, output_tokens=371
|
|
18:08:47,358 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:47,369 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.546999999321997. input_tokens=34, output_tokens=226
|
|
18:08:47,429 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:47,432 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.1409999998286366. input_tokens=1916, output_tokens=5
|
|
18:08:47,999 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:48,1 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.4059999994933605. input_tokens=34, output_tokens=127
|
|
18:08:49,704 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:49,707 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.906000000424683. input_tokens=2935, output_tokens=393
|
|
18:08:49,942 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:49,943 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.42200000025332. input_tokens=34, output_tokens=228
|
|
18:08:50,215 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:50,216 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 20.234000000171363. input_tokens=34, output_tokens=336
|
|
18:08:52,538 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:52,539 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 6.609000000171363. input_tokens=2936, output_tokens=122
|
|
18:08:53,296 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:53,298 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.15599999949336. input_tokens=34, output_tokens=146
|
|
18:08:53,390 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:53,392 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.89100000075996. input_tokens=2937, output_tokens=540
|
|
18:08:54,361 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:54,361 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.156999999657273. input_tokens=34, output_tokens=246
|
|
18:08:54,576 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:54,577 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.359000000171363. input_tokens=2937, output_tokens=255
|
|
18:08:56,103 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:56,117 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.07799999974668. input_tokens=2936, output_tokens=514
|
|
18:08:56,336 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:56,339 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.25. input_tokens=2936, output_tokens=581
|
|
18:08:56,912 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:56,914 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.95299999974668. input_tokens=2881, output_tokens=615
|
|
18:08:56,981 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:56,984 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.953999999910593. input_tokens=34, output_tokens=388
|
|
18:08:58,237 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:58,239 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.235000000335276. input_tokens=2936, output_tokens=428
|
|
18:08:58,421 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:58,424 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.843999999575317. input_tokens=2936, output_tokens=352
|
|
18:08:59,257 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:59,259 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.26600000075996. input_tokens=2937, output_tokens=291
|
|
18:08:59,811 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:08:59,815 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.781999999657273. input_tokens=34, output_tokens=243
|
|
18:09:00,401 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:00,401 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.781000000424683. input_tokens=2936, output_tokens=296
|
|
18:09:00,961 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:00,962 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.67200000025332. input_tokens=2936, output_tokens=141
|
|
18:09:01,346 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:01,347 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.17200000025332. input_tokens=2936, output_tokens=483
|
|
18:09:02,497 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:02,498 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.061999999918044. input_tokens=2937, output_tokens=283
|
|
18:09:02,907 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:02,908 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.968000000342727. input_tokens=2936, output_tokens=716
|
|
18:09:03,703 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:03,705 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.75. input_tokens=2938, output_tokens=282
|
|
18:09:04,106 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:04,110 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.563000000081956. input_tokens=34, output_tokens=226
|
|
18:09:05,113 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:05,129 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.421999999321997. input_tokens=2936, output_tokens=325
|
|
18:09:05,267 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:05,270 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.656000000424683. input_tokens=34, output_tokens=371
|
|
18:09:05,637 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:05,639 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.828000000678003. input_tokens=34, output_tokens=116
|
|
18:09:06,386 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:06,390 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.0. input_tokens=2937, output_tokens=416
|
|
18:09:06,423 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:06,426 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 32.51599999982864. input_tokens=2936, output_tokens=624
|
|
18:09:07,55 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:07,57 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.688000000081956. input_tokens=2936, output_tokens=563
|
|
18:09:07,665 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:07,667 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.311999999918044. input_tokens=2937, output_tokens=308
|
|
18:09:08,753 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:08,756 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.531000000424683. input_tokens=2937, output_tokens=639
|
|
18:09:08,941 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:08,944 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.686999999918044. input_tokens=34, output_tokens=226
|
|
18:09:09,352 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:09,352 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.718999999575317. input_tokens=34, output_tokens=97
|
|
18:09:09,652 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:09,658 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.95299999974668. input_tokens=34, output_tokens=165
|
|
18:09:11,77 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:11,81 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.5. input_tokens=2936, output_tokens=462
|
|
18:09:11,416 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:11,418 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.07799999974668. input_tokens=2935, output_tokens=560
|
|
18:09:12,80 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:12,82 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.406000000424683. input_tokens=2663, output_tokens=122
|
|
18:09:12,232 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:12,234 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.328999999910593. input_tokens=2937, output_tokens=279
|
|
18:09:12,522 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:12,530 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.109000000171363. input_tokens=2032, output_tokens=396
|
|
18:09:12,843 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:12,850 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.453999999910593. input_tokens=2936, output_tokens=291
|
|
18:09:13,412 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:13,413 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.29700000025332. input_tokens=2936, output_tokens=526
|
|
18:09:14,273 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:14,276 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.296000000089407. input_tokens=2936, output_tokens=421
|
|
18:09:14,402 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:14,404 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.281000000424683. input_tokens=34, output_tokens=198
|
|
18:09:15,291 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:15,297 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.07799999974668. input_tokens=2936, output_tokens=570
|
|
18:09:15,686 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:15,688 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.6099999994039536. input_tokens=34, output_tokens=86
|
|
18:09:15,754 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:15,756 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.42200000025332. input_tokens=2935, output_tokens=544
|
|
18:09:15,893 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:15,897 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.843000000342727. input_tokens=34, output_tokens=234
|
|
18:09:15,994 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:15,996 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.046999999321997. input_tokens=2936, output_tokens=215
|
|
18:09:16,844 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:16,845 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.8440000005066395. input_tokens=2937, output_tokens=5
|
|
18:09:18,41 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:18,43 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.625. input_tokens=2937, output_tokens=340
|
|
18:09:18,70 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:18,173 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.936999999918044. input_tokens=34, output_tokens=158
|
|
18:09:18,305 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:18,307 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.656999999657273. input_tokens=34, output_tokens=191
|
|
18:09:18,411 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:18,413 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.17200000025332. input_tokens=2936, output_tokens=596
|
|
18:09:19,4 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:19,6 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 6.938000000081956. input_tokens=2937, output_tokens=165
|
|
18:09:19,487 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:19,488 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.125. input_tokens=34, output_tokens=240
|
|
18:09:19,746 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:19,747 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.843999999575317. input_tokens=34, output_tokens=108
|
|
18:09:20,652 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:20,654 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.686999999918044. input_tokens=2936, output_tokens=633
|
|
18:09:20,673 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:20,675 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.8279999997466803. input_tokens=34, output_tokens=93
|
|
18:09:21,297 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:21,303 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.046999999321997. input_tokens=2937, output_tokens=413
|
|
18:09:22,16 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:22,21 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.264999999664724. input_tokens=34, output_tokens=377
|
|
18:09:22,240 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:22,242 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.843999999575317. input_tokens=34, output_tokens=215
|
|
18:09:22,383 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:22,386 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.860000000335276. input_tokens=34, output_tokens=250
|
|
18:09:23,126 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:23,128 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.375. input_tokens=34, output_tokens=93
|
|
18:09:23,261 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:23,266 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.875. input_tokens=34, output_tokens=5
|
|
18:09:23,830 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:23,831 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.484000000171363. input_tokens=2936, output_tokens=698
|
|
18:09:24,348 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:24,349 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.063000000081956. input_tokens=34, output_tokens=207
|
|
18:09:25,75 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,75 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.656000000424683. input_tokens=34, output_tokens=206
|
|
18:09:25,90 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,93 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.781000000424683. input_tokens=34, output_tokens=221
|
|
18:09:25,215 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,217 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.734000000171363. input_tokens=34, output_tokens=155
|
|
18:09:25,326 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,328 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.936999999918044. input_tokens=2937, output_tokens=551
|
|
18:09:25,374 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,377 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.093999999575317. input_tokens=34, output_tokens=141
|
|
18:09:25,394 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:25,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.531000000424683. input_tokens=34, output_tokens=327
|
|
18:09:26,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:26,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.2029999997466803. input_tokens=1959, output_tokens=5
|
|
18:09:26,346 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:26,347 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.32799999974668. input_tokens=34, output_tokens=198
|
|
18:09:26,578 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:26,578 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.328000000678003. input_tokens=34, output_tokens=95
|
|
18:09:26,938 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:26,939 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 30.015999999828637. input_tokens=2936, output_tokens=764
|
|
18:09:27,602 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:27,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.5. input_tokens=2936, output_tokens=605
|
|
18:09:27,859 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:27,860 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.593999999575317. input_tokens=34, output_tokens=102
|
|
18:09:28,541 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:28,543 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.515999999828637. input_tokens=34, output_tokens=190
|
|
18:09:28,581 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:28,583 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.765999999828637. input_tokens=34, output_tokens=130
|
|
18:09:29,54 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:29,56 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.6880000000819564. input_tokens=34, output_tokens=101
|
|
18:09:29,79 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:29,81 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.8590000001713634. input_tokens=34, output_tokens=107
|
|
18:09:29,206 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:29,209 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.140999999828637. input_tokens=34, output_tokens=107
|
|
18:09:29,224 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:29,230 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.735000000335276. input_tokens=2936, output_tokens=689
|
|
18:09:30,287 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:30,300 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.25. input_tokens=2937, output_tokens=412
|
|
18:09:30,888 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:30,895 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.203000000678003. input_tokens=34, output_tokens=488
|
|
18:09:30,946 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:30,951 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.593000000342727. input_tokens=34, output_tokens=165
|
|
18:09:31,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:31,184 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.3279999997466803. input_tokens=34, output_tokens=71
|
|
18:09:32,550 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:32,552 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.9849999994039536. input_tokens=34, output_tokens=158
|
|
18:09:32,723 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:32,726 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.063000000081956. input_tokens=2934, output_tokens=372
|
|
18:09:33,181 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:33,187 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.063000000081956. input_tokens=34, output_tokens=280
|
|
18:09:34,500 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:34,502 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 23.07799999974668. input_tokens=34, output_tokens=468
|
|
18:09:35,329 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:35,331 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.984000000171363. input_tokens=34, output_tokens=202
|
|
18:09:35,791 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:35,792 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.561999999918044. input_tokens=34, output_tokens=178
|
|
18:09:36,491 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:36,495 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.093999999575317. input_tokens=2936, output_tokens=262
|
|
18:09:36,501 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:36,503 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.20299999974668. input_tokens=34, output_tokens=279
|
|
18:09:37,75 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:37,77 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.531000000424683. input_tokens=34, output_tokens=246
|
|
18:09:37,487 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:37,489 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.390999999828637. input_tokens=34, output_tokens=226
|
|
18:09:38,121 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:38,126 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.796999999321997. input_tokens=2936, output_tokens=397
|
|
18:09:38,820 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:38,822 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.406000000424683. input_tokens=2936, output_tokens=747
|
|
18:09:38,910 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:38,912 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.265999999828637. input_tokens=2936, output_tokens=412
|
|
18:09:39,187 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:39,188 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.234999999403954. input_tokens=2936, output_tokens=227
|
|
18:09:40,167 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:40,169 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.95299999974668. input_tokens=2937, output_tokens=313
|
|
18:09:40,185 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:40,188 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.45299999974668. input_tokens=34, output_tokens=202
|
|
18:09:40,471 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:40,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.57799999974668. input_tokens=2937, output_tokens=282
|
|
18:09:40,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:40,571 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.390000000596046. input_tokens=2938, output_tokens=268
|
|
18:09:40,992 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:40,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.8279999997466803. input_tokens=1766, output_tokens=5
|
|
18:09:41,461 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:41,461 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.8909999998286366. input_tokens=2936, output_tokens=5
|
|
18:09:42,500 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:42,502 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.0. input_tokens=34, output_tokens=156
|
|
18:09:43,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:43,169 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.0. input_tokens=2935, output_tokens=713
|
|
18:09:43,310 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:43,313 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.70299999974668. input_tokens=2936, output_tokens=461
|
|
18:09:43,447 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:43,450 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.390000000596046. input_tokens=2936, output_tokens=432
|
|
18:09:45,244 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:45,246 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.921999999321997. input_tokens=2937, output_tokens=291
|
|
18:09:45,721 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:45,723 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.125. input_tokens=2826, output_tokens=436
|
|
18:09:46,172 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:46,176 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.25. input_tokens=34, output_tokens=193
|
|
18:09:46,620 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:46,622 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.061999999918044. input_tokens=2937, output_tokens=413
|
|
18:09:46,846 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:46,848 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.359000000171363. input_tokens=34, output_tokens=188
|
|
18:09:47,556 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:47,559 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.265999999828637. input_tokens=2936, output_tokens=510
|
|
18:09:47,800 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:47,803 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.5. input_tokens=2935, output_tokens=789
|
|
18:09:47,833 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:47,840 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.390999999828637. input_tokens=2936, output_tokens=76
|
|
18:09:47,915 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:47,919 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.42200000025332. input_tokens=2936, output_tokens=370
|
|
18:09:48,29 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:48,32 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.531000000424683. input_tokens=2936, output_tokens=326
|
|
18:09:48,174 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:48,176 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.67200000025332. input_tokens=34, output_tokens=151
|
|
18:09:48,366 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:48,367 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.186999999918044. input_tokens=34, output_tokens=249
|
|
18:09:48,839 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:48,842 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.656000000424683. input_tokens=2936, output_tokens=440
|
|
18:09:49,109 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:49,111 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.9380000000819564. input_tokens=2476, output_tokens=5
|
|
18:09:49,443 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:49,445 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.8900000005960464. input_tokens=2451, output_tokens=5
|
|
18:09:49,599 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:49,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.578999999910593. input_tokens=2936, output_tokens=5
|
|
18:09:50,102 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:50,104 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.281999999657273. input_tokens=2865, output_tokens=262
|
|
18:09:50,699 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:50,701 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.906000000424683. input_tokens=2936, output_tokens=259
|
|
18:09:51,509 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:51,511 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.6720000002533197. input_tokens=34, output_tokens=91
|
|
18:09:51,763 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:51,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.0630000000819564. input_tokens=34, output_tokens=5
|
|
18:09:52,397 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:52,398 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.093000000342727. input_tokens=2936, output_tokens=237
|
|
18:09:52,514 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:52,514 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:52,516 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.0. input_tokens=34, output_tokens=5
|
|
18:09:52,517 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.39100000075996. input_tokens=2937, output_tokens=345
|
|
18:09:52,576 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:52,578 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.9680000003427267. input_tokens=34, output_tokens=5
|
|
18:09:54,238 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:54,240 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.875. input_tokens=34, output_tokens=108
|
|
18:09:54,284 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:54,286 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.359000000171363. input_tokens=2938, output_tokens=800
|
|
18:09:55,929 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:55,930 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.859999999403954. input_tokens=2937, output_tokens=542
|
|
18:09:56,434 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:56,437 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.9069999996572733. input_tokens=34, output_tokens=96
|
|
18:09:58,229 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:58,230 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.656999999657273. input_tokens=34, output_tokens=113
|
|
18:09:58,278 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:58,280 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.561999999918044. input_tokens=2937, output_tokens=375
|
|
18:09:58,308 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:58,311 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.015999999828637. input_tokens=34, output_tokens=86
|
|
18:09:58,372 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:58,374 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.375. input_tokens=2936, output_tokens=386
|
|
18:09:58,849 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:58,851 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.093999999575317. input_tokens=34, output_tokens=125
|
|
18:09:59,400 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:59,402 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.234000000171363. input_tokens=34, output_tokens=374
|
|
18:09:59,693 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:59,695 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.234000000171363. input_tokens=2936, output_tokens=474
|
|
18:09:59,705 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:59,707 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.453000000678003. input_tokens=2937, output_tokens=298
|
|
18:09:59,840 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:09:59,842 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.390999999828637. input_tokens=34, output_tokens=222
|
|
18:10:00,765 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:00,768 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.14100000075996. input_tokens=2936, output_tokens=422
|
|
18:10:00,838 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:00,839 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.906000000424683. input_tokens=34, output_tokens=84
|
|
18:10:01,11 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:01,15 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 45.25. input_tokens=34, output_tokens=1190
|
|
18:10:02,420 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:02,423 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.90599999949336. input_tokens=2937, output_tokens=269
|
|
18:10:03,174 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:03,176 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.765999999828637. input_tokens=34, output_tokens=240
|
|
18:10:03,245 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:03,249 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.139999999664724. input_tokens=34, output_tokens=367
|
|
18:10:04,373 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:04,374 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.561999999918044. input_tokens=2936, output_tokens=390
|
|
18:10:04,795 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:04,796 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.093999999575317. input_tokens=34, output_tokens=87
|
|
18:10:05,159 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:05,161 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.843000000342727. input_tokens=34, output_tokens=134
|
|
18:10:06,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:06,471 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.031000000424683. input_tokens=2936, output_tokens=243
|
|
18:10:06,475 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:06,477 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.3130000000819564. input_tokens=34, output_tokens=79
|
|
18:10:06,666 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:06,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.9059999994933605. input_tokens=34, output_tokens=152
|
|
18:10:07,297 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:07,299 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.82799999974668. input_tokens=2935, output_tokens=696
|
|
18:10:07,588 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:07,594 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.313000000081956. input_tokens=2937, output_tokens=254
|
|
18:10:08,292 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:08,293 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.45299999974668. input_tokens=2937, output_tokens=243
|
|
18:10:09,17 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,18 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.781000000424683. input_tokens=34, output_tokens=298
|
|
18:10:09,110 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,112 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.9380000000819564. input_tokens=34, output_tokens=94
|
|
18:10:09,464 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,470 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.625. input_tokens=34, output_tokens=197
|
|
18:10:09,475 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,477 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.063000000081956. input_tokens=34, output_tokens=132
|
|
18:10:09,533 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,534 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.686999999918044. input_tokens=2936, output_tokens=577
|
|
18:10:09,734 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,736 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.625. input_tokens=2936, output_tokens=557
|
|
18:10:09,902 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:09,903 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.889999999664724. input_tokens=34, output_tokens=242
|
|
18:10:11,37 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:11,39 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 4.375. input_tokens=2936, output_tokens=111
|
|
18:10:11,121 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:11,122 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.421999999321997. input_tokens=2936, output_tokens=336
|
|
18:10:11,793 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:11,795 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 28.625. input_tokens=34, output_tokens=744
|
|
18:10:12,92 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:12,94 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.9690000005066395. input_tokens=34, output_tokens=5
|
|
18:10:12,489 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:12,490 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.093999999575317. input_tokens=34, output_tokens=359
|
|
18:10:12,977 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:12,979 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.9380000000819564. input_tokens=34, output_tokens=5
|
|
18:10:14,88 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:14,89 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.71900000050664. input_tokens=2936, output_tokens=271
|
|
18:10:14,830 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:14,833 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.25. input_tokens=34, output_tokens=197
|
|
18:10:14,987 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:14,989 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.8909999998286366. input_tokens=34, output_tokens=67
|
|
18:10:15,160 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:15,162 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.70299999974668. input_tokens=34, output_tokens=201
|
|
18:10:15,700 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:15,702 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.781000000424683. input_tokens=2937, output_tokens=822
|
|
18:10:15,852 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:15,853 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.563000000081956. input_tokens=34, output_tokens=198
|
|
18:10:16,25 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:16,31 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.5. input_tokens=34, output_tokens=98
|
|
18:10:16,92 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:16,95 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.84400000050664. input_tokens=34, output_tokens=262
|
|
18:10:16,502 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:16,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.656000000424683. input_tokens=34, output_tokens=472
|
|
18:10:16,525 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:16,538 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 27.70299999974668. input_tokens=34, output_tokens=735
|
|
18:10:16,978 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:16,980 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 36.79700000025332. input_tokens=2936, output_tokens=1034
|
|
18:10:17,79 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:17,84 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.593000000342727. input_tokens=2936, output_tokens=275
|
|
18:10:17,866 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:17,868 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.8899999996647239. input_tokens=34, output_tokens=5
|
|
18:10:18,189 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:18,191 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.70299999974668. input_tokens=2710, output_tokens=235
|
|
18:10:18,968 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:18,970 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.5. input_tokens=34, output_tokens=230
|
|
18:10:19,167 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:19,168 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.0. input_tokens=34, output_tokens=69
|
|
18:10:19,763 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:19,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.67200000025332. input_tokens=34, output_tokens=138
|
|
18:10:19,941 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:19,942 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.9069999996572733. input_tokens=34, output_tokens=93
|
|
18:10:20,150 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:20,152 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 25.906000000424683. input_tokens=34, output_tokens=559
|
|
18:10:20,207 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:20,208 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.186999999918044. input_tokens=2936, output_tokens=335
|
|
18:10:20,636 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:20,638 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.34400000050664. input_tokens=2936, output_tokens=434
|
|
18:10:21,447 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:21,453 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.718000000342727. input_tokens=2936, output_tokens=256
|
|
18:10:21,652 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:21,653 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.109000000171363. input_tokens=34, output_tokens=136
|
|
18:10:23,390 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:23,392 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.29700000025332. input_tokens=34, output_tokens=173
|
|
18:10:23,405 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:23,406 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:23,408 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.609000000171363. input_tokens=2936, output_tokens=436
|
|
18:10:23,410 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.296000000089407. input_tokens=2935, output_tokens=338
|
|
18:10:23,850 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:23,851 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.6409999998286366. input_tokens=34, output_tokens=84
|
|
18:10:24,234 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:24,235 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.718999999575317. input_tokens=2935, output_tokens=199
|
|
18:10:25,481 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:25,483 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 27.110000000335276. input_tokens=34, output_tokens=789
|
|
18:10:25,564 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:25,568 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.9220000002533197. input_tokens=2463, output_tokens=97
|
|
18:10:25,894 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:25,896 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.046000000089407. input_tokens=2936, output_tokens=265
|
|
18:10:26,386 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:26,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.45299999974668. input_tokens=34, output_tokens=172
|
|
18:10:26,464 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:26,466 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.82799999974668. input_tokens=34, output_tokens=151
|
|
18:10:27,367 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:27,369 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.9059999994933605. input_tokens=2856, output_tokens=5
|
|
18:10:27,585 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:27,588 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.17200000025332. input_tokens=34, output_tokens=107
|
|
18:10:28,64 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:28,68 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.203000000678003. input_tokens=2937, output_tokens=277
|
|
18:10:28,171 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:28,173 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.468999999575317. input_tokens=2392, output_tokens=356
|
|
18:10:28,253 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:28,257 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.281000000424683. input_tokens=2936, output_tokens=369
|
|
18:10:28,395 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:28,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.031000000424683. input_tokens=1754, output_tokens=5
|
|
18:10:30,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:30,290 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.8909999998286366. input_tokens=2936, output_tokens=89
|
|
18:10:30,983 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:30,986 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.79700000025332. input_tokens=2936, output_tokens=346
|
|
18:10:31,78 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:31,79 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.906000000424683. input_tokens=2938, output_tokens=356
|
|
18:10:31,293 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:31,295 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.53099999949336. input_tokens=34, output_tokens=320
|
|
18:10:31,761 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:31,769 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.860000000335276. input_tokens=2935, output_tokens=648
|
|
18:10:31,772 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:31,775 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.20299999974668. input_tokens=34, output_tokens=145
|
|
18:10:31,803 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:31,805 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.843999999575317. input_tokens=2936, output_tokens=331
|
|
18:10:32,533 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:32,535 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.2340000001713634. input_tokens=1828, output_tokens=5
|
|
18:10:32,697 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:32,763 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.67200000025332. input_tokens=2936, output_tokens=492
|
|
18:10:33,204 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:33,206 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.703000000678003. input_tokens=2936, output_tokens=559
|
|
18:10:33,489 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:33,491 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.9059999994933605. input_tokens=2936, output_tokens=168
|
|
18:10:34,406 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:34,409 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.32799999974668. input_tokens=34, output_tokens=205
|
|
18:10:35,84 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:35,86 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.188000000081956. input_tokens=2936, output_tokens=272
|
|
18:10:35,529 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:35,531 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.234000000171363. input_tokens=2936, output_tokens=134
|
|
18:10:35,652 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:35,654 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.484000000171363. input_tokens=34, output_tokens=208
|
|
18:10:35,718 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:35,719 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.313000000081956. input_tokens=34, output_tokens=142
|
|
18:10:35,850 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:35,852 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.3130000000819564. input_tokens=34, output_tokens=96
|
|
18:10:36,676 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:36,677 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.921999999321997. input_tokens=2936, output_tokens=118
|
|
18:10:36,769 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:36,772 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.0619999999180436. input_tokens=34, output_tokens=5
|
|
18:10:36,821 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:36,832 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.061999999918044. input_tokens=2935, output_tokens=136
|
|
18:10:37,352 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:37,354 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.906999999657273. input_tokens=2935, output_tokens=489
|
|
18:10:37,600 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:37,602 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.453999999910593. input_tokens=2937, output_tokens=523
|
|
18:10:37,627 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:37,630 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.21900000050664. input_tokens=2936, output_tokens=419
|
|
18:10:37,918 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:37,923 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.0619999999180436. input_tokens=34, output_tokens=46
|
|
18:10:38,461 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:38,466 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.8279999997466803. input_tokens=34, output_tokens=5
|
|
18:10:38,903 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:38,905 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.9840000001713634. input_tokens=34, output_tokens=5
|
|
18:10:40,482 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:40,485 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.640999999828637. input_tokens=2936, output_tokens=855
|
|
18:10:41,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:41,321 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.515000000596046. input_tokens=34, output_tokens=246
|
|
18:10:41,544 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:41,545 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.140999999828637. input_tokens=34, output_tokens=208
|
|
18:10:41,782 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:41,782 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.171000000089407. input_tokens=34, output_tokens=45
|
|
18:10:41,836 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:41,842 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.156000000424683. input_tokens=2567, output_tokens=161
|
|
18:10:42,535 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:42,537 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.15599999949336. input_tokens=2936, output_tokens=450
|
|
18:10:42,812 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:42,813 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.546999999321997. input_tokens=2935, output_tokens=472
|
|
18:10:43,90 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:43,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.625. input_tokens=34, output_tokens=74
|
|
18:10:43,519 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:43,526 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.734000000171363. input_tokens=2937, output_tokens=896
|
|
18:10:44,754 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:44,760 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.26600000075996. input_tokens=34, output_tokens=310
|
|
18:10:45,125 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,126 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.639999999664724. input_tokens=34, output_tokens=126
|
|
18:10:45,182 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,184 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,186 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.656999999657273. input_tokens=34, output_tokens=320
|
|
18:10:45,192 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.281999999657273. input_tokens=34, output_tokens=148
|
|
18:10:45,394 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,397 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.171000000089407. input_tokens=2937, output_tokens=635
|
|
18:10:45,532 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,534 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 32.54600000008941. input_tokens=2935, output_tokens=878
|
|
18:10:45,619 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,620 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.764999999664724. input_tokens=2938, output_tokens=612
|
|
18:10:45,940 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:45,942 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.156999999657273. input_tokens=34, output_tokens=417
|
|
18:10:46,345 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:46,347 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.8130000000819564. input_tokens=1742, output_tokens=5
|
|
18:10:46,600 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:46,601 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.953999999910593. input_tokens=2935, output_tokens=356
|
|
18:10:47,27 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:47,28 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.82799999974668. input_tokens=2936, output_tokens=357
|
|
18:10:49,212 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:49,213 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.140999999828637. input_tokens=2928, output_tokens=464
|
|
18:10:49,232 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:49,234 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.25. input_tokens=2936, output_tokens=388
|
|
18:10:49,333 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:49,335 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.25. input_tokens=34, output_tokens=158
|
|
18:10:49,662 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:49,666 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.859000000171363. input_tokens=34, output_tokens=189
|
|
18:10:49,672 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:49,674 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.0619999999180436. input_tokens=34, output_tokens=58
|
|
18:10:50,133 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:50,136 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.656000000424683. input_tokens=2936, output_tokens=610
|
|
18:10:50,490 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:50,492 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.733999999240041. input_tokens=34, output_tokens=158
|
|
18:10:50,529 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:50,533 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.20299999974668. input_tokens=34, output_tokens=196
|
|
18:10:50,872 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:50,875 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.093999999575317. input_tokens=34, output_tokens=277
|
|
18:10:52,794 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:52,797 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.593999999575317. input_tokens=34, output_tokens=169
|
|
18:10:53,22 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:53,24 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.484000000171363. input_tokens=34, output_tokens=204
|
|
18:10:53,136 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:53,138 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.313000000081956. input_tokens=34, output_tokens=551
|
|
18:10:53,759 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:53,760 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.235000000335276. input_tokens=34, output_tokens=70
|
|
18:10:53,987 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:53,988 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.953999999910593. input_tokens=34, output_tokens=5
|
|
18:10:54,358 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:54,361 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.015999999828637. input_tokens=34, output_tokens=234
|
|
18:10:55,210 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:55,210 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.0. input_tokens=34, output_tokens=97
|
|
18:10:55,427 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:55,429 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.563000000081956. input_tokens=34, output_tokens=137
|
|
18:10:56,121 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:56,125 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.171999999321997. input_tokens=34, output_tokens=299
|
|
18:10:56,398 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:56,401 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.561999999918044. input_tokens=2937, output_tokens=327
|
|
18:10:56,643 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:56,643 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.5. input_tokens=34, output_tokens=103
|
|
18:10:56,673 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:56,675 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.436999999918044. input_tokens=34, output_tokens=234
|
|
18:10:57,314 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:57,315 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.688000000081956. input_tokens=34, output_tokens=253
|
|
18:10:57,335 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:57,336 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.3590000001713634. input_tokens=34, output_tokens=91
|
|
18:10:57,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:57,687 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.5. input_tokens=34, output_tokens=281
|
|
18:10:58,339 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:58,350 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.578999999910593. input_tokens=2935, output_tokens=367
|
|
18:10:58,870 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:58,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.343999999575317. input_tokens=34, output_tokens=272
|
|
18:10:59,211 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:59,212 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.54700000025332. input_tokens=34, output_tokens=201
|
|
18:10:59,452 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:59,455 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.906000000424683. input_tokens=2936, output_tokens=434
|
|
18:10:59,807 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:59,812 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.45299999974668. input_tokens=2936, output_tokens=398
|
|
18:10:59,833 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:10:59,836 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.813000000081956. input_tokens=34, output_tokens=240
|
|
18:11:00,296 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:00,299 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.17200000025332. input_tokens=2936, output_tokens=356
|
|
18:11:00,305 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:00,307 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.640999999828637. input_tokens=34, output_tokens=186
|
|
18:11:00,424 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:00,426 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.07799999974668. input_tokens=34, output_tokens=353
|
|
18:11:00,602 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:00,604 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.515999999828637. input_tokens=2937, output_tokens=791
|
|
18:11:01,35 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:01,38 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.54700000025332. input_tokens=34, output_tokens=288
|
|
18:11:01,190 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:01,193 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.9060000004246831. input_tokens=34, output_tokens=5
|
|
18:11:02,998 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:03,2 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.593999999575317. input_tokens=34, output_tokens=176
|
|
18:11:03,369 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:03,371 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.5619999999180436. input_tokens=34, output_tokens=82
|
|
18:11:04,568 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:04,574 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.811999999918044. input_tokens=34, output_tokens=176
|
|
18:11:05,301 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:05,319 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.92200000025332. input_tokens=2842, output_tokens=490
|
|
18:11:05,679 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:05,682 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.32799999974668. input_tokens=34, output_tokens=208
|
|
18:11:06,961 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:06,963 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.17200000025332. input_tokens=34, output_tokens=474
|
|
18:11:07,955 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:07,957 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.811999999918044. input_tokens=34, output_tokens=553
|
|
18:11:09,442 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:09,443 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.0. input_tokens=34, output_tokens=239
|
|
18:11:09,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:09,858 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.531999999657273. input_tokens=34, output_tokens=107
|
|
18:11:13,338 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:13,339 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.734000000171363. input_tokens=34, output_tokens=320
|
|
18:11:13,358 datashaper.workflow.workflow INFO executing verb merge_graphs
|
|
18:11:13,529 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_extracted_entities.parquet
|
|
18:11:13,712 graphrag.index.run.workflow INFO dependencies for create_summarized_entities: ['create_base_extracted_entities']
|
|
18:11:13,713 graphrag.utils.storage INFO read table from storage: create_base_extracted_entities.parquet
|
|
18:11:13,732 datashaper.workflow.workflow INFO executing verb summarize_descriptions
|
|
18:11:17,219 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:17,221 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.3590000001713634. input_tokens=162, output_tokens=71
|
|
18:11:17,292 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:17,296 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.4369999999180436. input_tokens=176, output_tokens=84
|
|
18:11:17,621 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:17,622 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.764999999664724. input_tokens=146, output_tokens=48
|
|
18:11:18,77 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:18,78 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.17200000025332. input_tokens=194, output_tokens=103
|
|
18:11:18,220 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:18,222 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.313000000081956. input_tokens=163, output_tokens=65
|
|
18:11:18,586 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:18,588 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.7190000005066395. input_tokens=190, output_tokens=103
|
|
18:11:19,12 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:19,14 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.156000000424683. input_tokens=212, output_tokens=132
|
|
18:11:19,687 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:19,692 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.875. input_tokens=310, output_tokens=178
|
|
18:11:20,547 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:20,550 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.718999999575317. input_tokens=241, output_tokens=146
|
|
18:11:22,149 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:22,151 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.859000000171363. input_tokens=148, output_tokens=58
|
|
18:11:22,250 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:22,255 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.171999999321997. input_tokens=179, output_tokens=81
|
|
18:11:22,260 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:22,262 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.34400000050664. input_tokens=200, output_tokens=191
|
|
18:11:22,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:22,686 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.468999999575317. input_tokens=191, output_tokens=127
|
|
18:11:23,721 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:23,722 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.890999999828637. input_tokens=237, output_tokens=181
|
|
18:11:24,197 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:24,199 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.29700000025332. input_tokens=290, output_tokens=192
|
|
18:11:24,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:24,290 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.2809999994933605. input_tokens=202, output_tokens=124
|
|
18:11:24,945 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:24,946 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.125. input_tokens=389, output_tokens=297
|
|
18:11:25,411 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:25,416 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.875. input_tokens=174, output_tokens=112
|
|
18:11:27,11 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:27,11 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.17200000025332. input_tokens=258, output_tokens=233
|
|
18:11:27,364 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:27,365 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6559999994933605. input_tokens=196, output_tokens=82
|
|
18:11:27,389 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:27,391 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.688000000081956. input_tokens=296, output_tokens=203
|
|
18:11:27,461 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:27,462 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:27,464 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.640999999828637. input_tokens=381, output_tokens=337
|
|
18:11:27,468 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.640999999828637. input_tokens=298, output_tokens=248
|
|
18:11:28,225 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:28,225 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.015999999828637. input_tokens=309, output_tokens=314
|
|
18:11:28,431 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:28,433 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.171999999321997. input_tokens=229, output_tokens=122
|
|
18:11:28,668 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:28,673 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.25. input_tokens=157, output_tokens=60
|
|
18:11:29,254 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:29,257 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.42200000025332. input_tokens=635, output_tokens=435
|
|
18:11:29,340 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:29,344 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.3279999997466803. input_tokens=148, output_tokens=38
|
|
18:11:29,701 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:29,703 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.811999999918044. input_tokens=1142, output_tokens=499
|
|
18:11:29,756 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:29,760 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.89100000075996. input_tokens=774, output_tokens=500
|
|
18:11:29,772 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:29,774 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.546000000089407. input_tokens=134, output_tokens=24
|
|
18:11:30,519 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:30,521 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 16.625. input_tokens=1134, output_tokens=500
|
|
18:11:30,651 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:30,652 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.968000000342727. input_tokens=251, output_tokens=246
|
|
18:11:30,806 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:30,808 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.656999999657273. input_tokens=227, output_tokens=156
|
|
18:11:31,549 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:31,553 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 17.671999999321997. input_tokens=943, output_tokens=500
|
|
18:11:31,766 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:31,768 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.391000000759959. input_tokens=162, output_tokens=101
|
|
18:11:32,57 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:32,59 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.593999999575317. input_tokens=144, output_tokens=50
|
|
18:11:32,84 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:32,88 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 18.25. input_tokens=560, output_tokens=498
|
|
18:11:32,440 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:32,442 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.234999999403954. input_tokens=230, output_tokens=199
|
|
18:11:32,762 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:32,763 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.328000000678003. input_tokens=149, output_tokens=65
|
|
18:11:33,514 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:33,516 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.75. input_tokens=159, output_tokens=93
|
|
18:11:33,537 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:33,539 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.07799999974668. input_tokens=212, output_tokens=141
|
|
18:11:33,891 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:33,893 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 20.01600000075996. input_tokens=556, output_tokens=476
|
|
18:11:34,25 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:34,30 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.32799999974668. input_tokens=196, output_tokens=87
|
|
18:11:34,836 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:34,838 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 20.984000000171363. input_tokens=1155, output_tokens=500
|
|
18:11:34,860 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:34,861 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.203999999910593. input_tokens=183, output_tokens=78
|
|
18:11:35,145 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,148 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.343000000342727. input_tokens=156, output_tokens=67
|
|
18:11:35,246 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,249 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.1869999999180436. input_tokens=163, output_tokens=73
|
|
18:11:35,531 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,535 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.186999999918044. input_tokens=223, output_tokens=185
|
|
18:11:35,567 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,568 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.311999999918044. input_tokens=211, output_tokens=122
|
|
18:11:35,667 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,675 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.0. input_tokens=228, output_tokens=222
|
|
18:11:35,886 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:35,889 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 22.04700000025332. input_tokens=2113, output_tokens=500
|
|
18:11:36,61 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:36,64 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.296999999321997. input_tokens=154, output_tokens=72
|
|
18:11:36,206 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:36,244 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.8119999999180436. input_tokens=163, output_tokens=84
|
|
18:11:37,354 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:37,356 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.265999999828637. input_tokens=171, output_tokens=126
|
|
18:11:37,393 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:37,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.8440000005066395. input_tokens=154, output_tokens=86
|
|
18:11:37,674 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:37,676 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.28099999949336. input_tokens=236, output_tokens=232
|
|
18:11:37,845 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:37,847 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.25. input_tokens=556, output_tokens=459
|
|
18:11:38,209 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:38,211 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.438000000081956. input_tokens=415, output_tokens=173
|
|
18:11:39,32 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:39,33 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 21.406000000424683. input_tokens=1282, output_tokens=500
|
|
18:11:39,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:39,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.0. input_tokens=384, output_tokens=378
|
|
18:11:39,532 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:39,532 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.014999999664724. input_tokens=157, output_tokens=121
|
|
18:11:39,699 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:39,700 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 17.436999999918044. input_tokens=542, output_tokens=446
|
|
18:11:40,405 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:40,406 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.343000000342727. input_tokens=204, output_tokens=99
|
|
18:11:40,766 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:40,767 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.610000000335276. input_tokens=197, output_tokens=134
|
|
18:11:41,75 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:41,77 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.04700000025332. input_tokens=256, output_tokens=172
|
|
18:11:41,366 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:41,369 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.60899999924004. input_tokens=218, output_tokens=152
|
|
18:11:41,429 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:41,430 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.188000000081956. input_tokens=191, output_tokens=126
|
|
18:11:41,640 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:41,641 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.75. input_tokens=205, output_tokens=167
|
|
18:11:41,997 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:41,998 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.10899999924004. input_tokens=325, output_tokens=222
|
|
18:11:42,653 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:42,655 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.984000000171363. input_tokens=170, output_tokens=93
|
|
18:11:42,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:42,782 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.375. input_tokens=164, output_tokens=106
|
|
18:11:43,70 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:43,72 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.375. input_tokens=147, output_tokens=67
|
|
18:11:43,160 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:43,162 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.875. input_tokens=163, output_tokens=58
|
|
18:11:43,635 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:43,636 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.42200000025332. input_tokens=173, output_tokens=82
|
|
18:11:43,712 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:43,714 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.859000000171363. input_tokens=229, output_tokens=193
|
|
18:11:44,5 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:44,9 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.76600000075996. input_tokens=219, output_tokens=211
|
|
18:11:44,160 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:44,162 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.7340000001713634. input_tokens=166, output_tokens=70
|
|
18:11:44,634 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:44,640 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.110000000335276. input_tokens=154, output_tokens=100
|
|
18:11:45,159 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:45,159 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.311999999918044. input_tokens=162, output_tokens=122
|
|
18:11:45,256 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:45,258 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 20.313000000081956. input_tokens=562, output_tokens=436
|
|
18:11:45,316 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:45,318 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.92200000025332. input_tokens=155, output_tokens=83
|
|
18:11:45,443 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:45,444 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.609000000171363. input_tokens=300, output_tokens=203
|
|
18:11:45,605 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:45,609 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.031999999657273. input_tokens=242, output_tokens=193
|
|
18:11:46,409 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:46,410 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.764999999664724. input_tokens=166, output_tokens=57
|
|
18:11:46,545 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:46,548 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.015999999828637. input_tokens=250, output_tokens=225
|
|
18:11:46,649 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:46,653 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.093000000342727. input_tokens=446, output_tokens=437
|
|
18:11:47,202 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:47,401 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:47,403 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.375. input_tokens=177, output_tokens=137
|
|
18:11:47,424 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:47,425 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.6559999994933605. input_tokens=163, output_tokens=90
|
|
18:11:47,435 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.265999999828637. input_tokens=179, output_tokens=68
|
|
18:11:47,791 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:47,792 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.625. input_tokens=166, output_tokens=58
|
|
18:11:48,74 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:48,75 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.9220000002533197. input_tokens=146, output_tokens=41
|
|
18:11:48,375 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:48,379 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.921999999321997. input_tokens=170, output_tokens=72
|
|
18:11:48,419 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:48,419 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.093999999575317. input_tokens=163, output_tokens=76
|
|
18:11:48,427 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:48,428 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.171999999321997. input_tokens=156, output_tokens=77
|
|
18:11:48,741 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:48,742 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.093999999575317. input_tokens=228, output_tokens=113
|
|
18:11:49,270 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:49,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.20299999974668. input_tokens=207, output_tokens=121
|
|
18:11:49,299 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:49,300 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.625. input_tokens=292, output_tokens=280
|
|
18:11:49,310 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:49,312 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.313000000081956. input_tokens=218, output_tokens=131
|
|
18:11:49,956 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:49,958 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.406000000424683. input_tokens=153, output_tokens=72
|
|
18:11:50,76 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:50,77 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.656000000424683. input_tokens=187, output_tokens=49
|
|
18:11:50,522 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:50,528 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.125. input_tokens=187, output_tokens=97
|
|
18:11:50,976 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:50,984 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.203999999910593. input_tokens=212, output_tokens=146
|
|
18:11:51,9 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,13 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.0. input_tokens=210, output_tokens=160
|
|
18:11:51,211 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,212 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.8130000000819564. input_tokens=152, output_tokens=87
|
|
18:11:51,235 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,237 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.515999999828637. input_tokens=239, output_tokens=189
|
|
18:11:51,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,350 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5630000000819564. input_tokens=181, output_tokens=71
|
|
18:11:51,350 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.281999999657273. input_tokens=253, output_tokens=164
|
|
18:11:51,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,480 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 20.953999999910593. input_tokens=406, output_tokens=428
|
|
18:11:51,773 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:51,776 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.171000000089407. input_tokens=182, output_tokens=150
|
|
18:11:52,225 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:52,229 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.593999999575317. input_tokens=343, output_tokens=226
|
|
18:11:52,332 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:52,334 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.96900000050664. input_tokens=222, output_tokens=265
|
|
18:11:52,606 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:52,607 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.953999999910593. input_tokens=186, output_tokens=125
|
|
18:11:52,623 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:52,625 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.186999999918044. input_tokens=199, output_tokens=103
|
|
18:11:52,774 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:52,775 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.359000000171363. input_tokens=188, output_tokens=94
|
|
18:11:52,998 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:53,0 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.639999999664724. input_tokens=288, output_tokens=316
|
|
18:11:53,609 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:53,612 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.17200000025332. input_tokens=192, output_tokens=71
|
|
18:11:53,689 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:53,693 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.3900000005960464. input_tokens=187, output_tokens=98
|
|
18:11:53,770 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:53,772 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.70299999974668. input_tokens=185, output_tokens=85
|
|
18:11:54,97 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:54,99 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.735000000335276. input_tokens=187, output_tokens=71
|
|
18:11:54,105 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:54,121 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.48399999924004. input_tokens=231, output_tokens=247
|
|
18:11:54,314 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:54,316 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.2349999994039536. input_tokens=216, output_tokens=91
|
|
18:11:54,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:54,802 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5779999997466803. input_tokens=189, output_tokens=81
|
|
18:11:54,951 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:54,953 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.42200000025332. input_tokens=220, output_tokens=93
|
|
18:11:55,332 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:55,335 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.04700000025332. input_tokens=208, output_tokens=120
|
|
18:11:55,606 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:55,611 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.25. input_tokens=195, output_tokens=109
|
|
18:11:55,829 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:55,830 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.859000000171363. input_tokens=253, output_tokens=109
|
|
18:11:56,50 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,53 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.313000000081956. input_tokens=229, output_tokens=181
|
|
18:11:56,116 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,119 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.343999999575317. input_tokens=176, output_tokens=77
|
|
18:11:56,519 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,526 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.296000000089407. input_tokens=186, output_tokens=77
|
|
18:11:56,531 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,533 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.921000000089407. input_tokens=170, output_tokens=87
|
|
18:11:56,558 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,560 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.281999999657273. input_tokens=238, output_tokens=158
|
|
18:11:56,861 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:56,864 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.640999999828637. input_tokens=246, output_tokens=178
|
|
18:11:57,483 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:57,488 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.0. input_tokens=207, output_tokens=138
|
|
18:11:57,600 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:57,643 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.625. input_tokens=207, output_tokens=153
|
|
18:11:57,853 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:57,855 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5470000002533197. input_tokens=193, output_tokens=82
|
|
18:11:57,879 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:57,880 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.0630000000819564. input_tokens=151, output_tokens=36
|
|
18:11:57,908 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:57,910 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.921000000089407. input_tokens=237, output_tokens=188
|
|
18:11:58,345 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:58,347 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.984000000171363. input_tokens=231, output_tokens=148
|
|
18:11:59,464 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:11:59,466 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.515999999828637. input_tokens=169, output_tokens=115
|
|
18:12:00,22 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:00,26 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5. input_tokens=196, output_tokens=76
|
|
18:12:00,277 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:00,280 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.936999999918044. input_tokens=217, output_tokens=136
|
|
18:12:00,709 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:00,710 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.375. input_tokens=223, output_tokens=167
|
|
18:12:00,714 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:00,715 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.5940000005066395. input_tokens=260, output_tokens=144
|
|
18:12:00,864 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:00,869 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.264999999664724. input_tokens=223, output_tokens=101
|
|
18:12:01,15 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:01,73 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.2650000005960464. input_tokens=216, output_tokens=156
|
|
18:12:01,287 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:01,289 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.234000000171363. input_tokens=169, output_tokens=70
|
|
18:12:01,317 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:01,319 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6869999999180436. input_tokens=218, output_tokens=84
|
|
18:12:01,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:01,868 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.968999999575317. input_tokens=165, output_tokens=86
|
|
18:12:02,160 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:02,162 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.390999999828637. input_tokens=272, output_tokens=251
|
|
18:12:02,620 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:02,624 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.061999999918044. input_tokens=231, output_tokens=151
|
|
18:12:02,795 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:02,797 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.311999999918044. input_tokens=221, output_tokens=133
|
|
18:12:03,56 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:03,58 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.938000000081956. input_tokens=307, output_tokens=179
|
|
18:12:03,824 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:03,825 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.953000000678003. input_tokens=248, output_tokens=132
|
|
18:12:05,306 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,314 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 12.688000000081956. input_tokens=533, output_tokens=375
|
|
18:12:05,333 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,342 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.563000000081956. input_tokens=255, output_tokens=280
|
|
18:12:05,369 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,370 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.671999999321997. input_tokens=437, output_tokens=355
|
|
18:12:05,604 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,607 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.735000000335276. input_tokens=173, output_tokens=80
|
|
18:12:05,618 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,619 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.32799999974668. input_tokens=165, output_tokens=90
|
|
18:12:05,685 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,690 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.796999999321997. input_tokens=169, output_tokens=67
|
|
18:12:05,992 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:05,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.718999999575317. input_tokens=196, output_tokens=103
|
|
18:12:06,281 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:06,283 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.20299999974668. input_tokens=235, output_tokens=138
|
|
18:12:06,349 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:06,351 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.5. input_tokens=263, output_tokens=157
|
|
18:12:06,777 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:06,783 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.061999999918044. input_tokens=268, output_tokens=173
|
|
18:12:06,854 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:06,856 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.140999999828637. input_tokens=252, output_tokens=147
|
|
18:12:07,370 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:07,372 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.343999999575317. input_tokens=378, output_tokens=181
|
|
18:12:08,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:08,174 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.561999999918044. input_tokens=491, output_tokens=363
|
|
18:12:08,393 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:08,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.281000000424683. input_tokens=439, output_tokens=317
|
|
18:12:09,385 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:09,389 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 16.39100000075996. input_tokens=420, output_tokens=374
|
|
18:12:09,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:09,801 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.32799999974668. input_tokens=352, output_tokens=258
|
|
18:12:12,659 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:12,662 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.32799999974668. input_tokens=323, output_tokens=387
|
|
18:12:14,159 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:14,160 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 17.625. input_tokens=537, output_tokens=424
|
|
18:12:14,211 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_summarized_entities.parquet
|
|
18:12:14,364 graphrag.index.run.workflow INFO dependencies for create_base_entity_graph: ['create_summarized_entities']
|
|
18:12:14,365 graphrag.utils.storage INFO read table from storage: create_summarized_entities.parquet
|
|
18:12:14,387 datashaper.workflow.workflow INFO executing verb cluster_graph
|
|
18:12:14,871 datashaper.workflow.workflow INFO executing verb select
|
|
18:12:14,875 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_entity_graph.parquet
|
|
18:12:15,60 graphrag.index.run.workflow INFO dependencies for create_final_entities: ['create_base_entity_graph']
|
|
18:12:15,67 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:12:15,100 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:12:15,298 datashaper.workflow.workflow INFO executing verb rename
|
|
18:12:15,308 datashaper.workflow.workflow INFO executing verb select
|
|
18:12:15,323 datashaper.workflow.workflow INFO executing verb dedupe
|
|
18:12:15,337 datashaper.workflow.workflow INFO executing verb rename
|
|
18:12:15,348 datashaper.workflow.workflow INFO executing verb filter
|
|
18:12:15,381 datashaper.workflow.workflow INFO executing verb text_split
|
|
18:12:15,400 datashaper.workflow.workflow INFO executing verb drop
|
|
18:12:15,412 datashaper.workflow.workflow INFO executing verb merge
|
|
18:12:15,580 datashaper.workflow.workflow INFO executing verb text_embed
|
|
18:12:15,581 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:12:15,591 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for text-embedding-3-small: TPM=0, RPM=0
|
|
18:12:15,591 graphrag.index.llm.load_llm INFO create concurrency limiter for text-embedding-3-small: 25
|
|
18:12:15,688 graphrag.index.verbs.text.embed.strategies.openai INFO embedding 628 inputs via 628 snippets using 40 batches. max_batch_size=16, max_tokens=8191
|
|
18:12:16,268 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,305 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,338 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,374 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,422 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,498 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,532 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,581 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,590 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8909999998286366. input_tokens=997, output_tokens=0
|
|
18:12:16,607 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,609 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,641 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,653 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,660 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,663 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,676 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,722 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,743 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0149999996647239. input_tokens=1382, output_tokens=0
|
|
18:12:16,800 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0779999997466803. input_tokens=755, output_tokens=0
|
|
18:12:16,818 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,835 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.1090000001713634. input_tokens=1734, output_tokens=0
|
|
18:12:16,865 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,869 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.1559999994933605. input_tokens=1482, output_tokens=0
|
|
18:12:16,923 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,934 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.218999999575317. input_tokens=1417, output_tokens=0
|
|
18:12:16,953 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:16,982 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.25. input_tokens=1211, output_tokens=0
|
|
18:12:17,8 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,14 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2660000007599592. input_tokens=782, output_tokens=0
|
|
18:12:17,43 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3279999997466803. input_tokens=893, output_tokens=0
|
|
18:12:17,88 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3590000001713634. input_tokens=823, output_tokens=0
|
|
18:12:17,120 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4219999993219972. input_tokens=359, output_tokens=0
|
|
18:12:17,142 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,143 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,151 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4529999997466803. input_tokens=3000, output_tokens=0
|
|
18:12:17,186 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,188 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4529999997466803. input_tokens=664, output_tokens=0
|
|
18:12:17,211 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4840000001713634. input_tokens=996, output_tokens=0
|
|
18:12:17,234 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,240 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5. input_tokens=495, output_tokens=0
|
|
18:12:17,265 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5470000002533197. input_tokens=1029, output_tokens=0
|
|
18:12:17,286 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.546000000089407. input_tokens=1204, output_tokens=0
|
|
18:12:17,307 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,308 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,334 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6090000001713634. input_tokens=954, output_tokens=0
|
|
18:12:17,349 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,354 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6409999998286366. input_tokens=1146, output_tokens=0
|
|
18:12:17,374 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6719999993219972. input_tokens=1108, output_tokens=0
|
|
18:12:17,396 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7029999997466803. input_tokens=3205, output_tokens=0
|
|
18:12:17,417 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6720000002533197. input_tokens=874, output_tokens=0
|
|
18:12:17,437 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7029999997466803. input_tokens=656, output_tokens=0
|
|
18:12:17,459 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,468 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7340000001713634. input_tokens=850, output_tokens=0
|
|
18:12:17,509 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,520 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.7180000003427267. input_tokens=655, output_tokens=0
|
|
18:12:17,577 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.8590000001713634. input_tokens=830, output_tokens=0
|
|
18:12:17,617 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,648 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,650 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.3430000003427267. input_tokens=53, output_tokens=0
|
|
18:12:17,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0619999999180436. input_tokens=605, output_tokens=0
|
|
18:12:17,686 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,691 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8130000000819564. input_tokens=755, output_tokens=0
|
|
18:12:17,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,862 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,928 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8130000000819564. input_tokens=566, output_tokens=0
|
|
18:12:17,949 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:17,998 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:18,2 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2339999992400408. input_tokens=385, output_tokens=0
|
|
18:12:18,24 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:12:18,74 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.875. input_tokens=798, output_tokens=0
|
|
18:12:18,126 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0619999999180436. input_tokens=512, output_tokens=0
|
|
18:12:18,227 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.375. input_tokens=500, output_tokens=0
|
|
18:12:18,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.218999999575317. input_tokens=476, output_tokens=0
|
|
18:12:18,275 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.046000000089407. input_tokens=509, output_tokens=0
|
|
18:12:18,341 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2029999997466803. input_tokens=495, output_tokens=0
|
|
18:12:18,404 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2340000001713634. input_tokens=416, output_tokens=0
|
|
18:12:18,422 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5309999994933605. input_tokens=918, output_tokens=0
|
|
18:12:18,570 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5780000006780028. input_tokens=623, output_tokens=0
|
|
18:12:18,614 datashaper.workflow.workflow INFO executing verb drop
|
|
18:12:18,627 datashaper.workflow.workflow INFO executing verb filter
|
|
18:12:18,659 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_entities.parquet
|
|
18:12:18,954 graphrag.index.run.workflow INFO dependencies for create_final_nodes: ['create_base_entity_graph']
|
|
18:12:18,955 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:12:19,0 datashaper.workflow.workflow INFO executing verb layout_graph
|
|
18:12:19,912 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:12:20,343 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:12:20,671 datashaper.workflow.workflow INFO executing verb filter
|
|
18:12:20,748 datashaper.workflow.workflow INFO executing verb drop
|
|
18:12:20,944 datashaper.workflow.workflow INFO executing verb select
|
|
18:12:20,960 datashaper.workflow.workflow INFO executing verb rename
|
|
18:12:20,984 datashaper.workflow.workflow INFO executing verb convert
|
|
18:12:21,46 datashaper.workflow.workflow INFO executing verb join
|
|
18:12:21,79 datashaper.workflow.workflow INFO executing verb rename
|
|
18:12:21,82 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_nodes.parquet
|
|
18:12:21,313 graphrag.index.run.workflow INFO dependencies for create_final_communities: ['create_base_entity_graph']
|
|
18:12:21,314 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:12:21,363 datashaper.workflow.workflow INFO executing verb create_final_communities
|
|
18:12:21,913 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_communities.parquet
|
|
18:12:22,133 graphrag.index.run.workflow INFO dependencies for create_final_relationships: ['create_final_nodes', 'create_base_entity_graph']
|
|
18:12:22,134 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:12:22,143 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:12:22,194 datashaper.workflow.workflow INFO executing verb create_final_relationships_pre_embedding
|
|
18:12:22,427 datashaper.workflow.workflow INFO executing verb create_final_relationships_post_embedding
|
|
18:12:22,437 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_relationships.parquet
|
|
18:12:22,638 graphrag.index.run.workflow INFO dependencies for create_final_text_units: ['create_final_relationships', 'create_base_text_units', 'create_final_entities']
|
|
18:12:22,639 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:12:22,644 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:12:22,650 graphrag.utils.storage INFO read table from storage: create_final_entities.parquet
|
|
18:12:22,731 datashaper.workflow.workflow INFO executing verb create_final_text_units_pre_embedding
|
|
18:12:22,775 datashaper.workflow.workflow INFO executing verb select
|
|
18:12:22,778 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_text_units.parquet
|
|
18:12:22,969 graphrag.index.run.workflow INFO dependencies for create_final_community_reports: ['create_final_relationships', 'create_final_nodes']
|
|
18:12:22,980 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:12:22,986 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:12:23,55 datashaper.workflow.workflow INFO executing verb prepare_community_reports_nodes
|
|
18:12:23,135 datashaper.workflow.workflow INFO executing verb prepare_community_reports_edges
|
|
18:12:23,178 datashaper.workflow.workflow INFO executing verb restore_community_hierarchy
|
|
18:12:23,210 datashaper.workflow.workflow INFO executing verb prepare_community_reports
|
|
18:12:23,212 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=3 => 628
|
|
18:12:23,248 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=2 => 628
|
|
18:12:23,334 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=1 => 628
|
|
18:12:23,567 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=0 => 628
|
|
18:12:23,799 datashaper.workflow.workflow INFO executing verb create_community_reports
|
|
18:12:48,721 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:48,724 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.890999999828637. input_tokens=5195, output_tokens=761
|
|
18:12:52,414 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:52,416 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.609000000171363. input_tokens=3518, output_tokens=755
|
|
18:12:57,458 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:12:57,460 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.64099999982864. input_tokens=2664, output_tokens=696
|
|
18:13:13,843 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:13,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 50.171999999322. input_tokens=4822, output_tokens=986
|
|
18:13:31,42 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:31,45 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.936999999918044. input_tokens=2236, output_tokens=354
|
|
18:13:32,103 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:32,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.985000000335276. input_tokens=2388, output_tokens=586
|
|
18:13:38,228 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:38,234 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.140999999828637. input_tokens=5474, output_tokens=819
|
|
18:13:40,969 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:40,971 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.875. input_tokens=2377, output_tokens=642
|
|
18:13:42,495 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:42,502 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.421999999321997. input_tokens=2213, output_tokens=687
|
|
18:13:43,247 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:43,255 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.20299999974668. input_tokens=5455, output_tokens=936
|
|
18:13:44,697 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:44,701 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:44,704 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.656000000424683. input_tokens=6604, output_tokens=894
|
|
18:13:44,708 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.625. input_tokens=4041, output_tokens=903
|
|
18:13:49,491 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:49,496 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 35.436999999918044. input_tokens=6266, output_tokens=901
|
|
18:13:50,194 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:13:50,207 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.078000000678. input_tokens=4453, output_tokens=867
|
|
18:14:08,647 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:08,648 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.234000000171363. input_tokens=2153, output_tokens=570
|
|
18:14:10,576 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:10,578 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.186999999918044. input_tokens=2149, output_tokens=654
|
|
18:14:11,265 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:11,267 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.985000000335276. input_tokens=2093, output_tokens=547
|
|
18:14:13,461 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:13,464 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.04700000025332. input_tokens=2222, output_tokens=505
|
|
18:14:16,960 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:16,962 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.563000000081956. input_tokens=2489, output_tokens=705
|
|
18:14:16,987 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:16,989 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.718999999575317. input_tokens=2325, output_tokens=635
|
|
18:14:18,319 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:18,322 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.936999999918044. input_tokens=2620, output_tokens=755
|
|
18:14:18,364 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:18,366 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.046999999321997. input_tokens=2133, output_tokens=583
|
|
18:14:19,505 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:19,507 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.26600000075996. input_tokens=2927, output_tokens=762
|
|
18:14:19,863 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:19,866 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.53099999949336. input_tokens=2317, output_tokens=761
|
|
18:14:19,940 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:19,942 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.640999999828637. input_tokens=6297, output_tokens=777
|
|
18:14:20,673 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:20,677 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.265999999828637. input_tokens=3992, output_tokens=793
|
|
18:14:20,738 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:20,740 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.389999999664724. input_tokens=2778, output_tokens=801
|
|
18:14:22,82 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:22,85 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.82799999974668. input_tokens=9779, output_tokens=770
|
|
18:14:22,162 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:22,166 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.82799999974668. input_tokens=6365, output_tokens=874
|
|
18:14:22,254 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:22,257 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 13.610000000335276. input_tokens=2062, output_tokens=439
|
|
18:14:24,170 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:24,172 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.85900000017136. input_tokens=2241, output_tokens=730
|
|
18:14:24,805 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:24,807 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.48499999940395. input_tokens=2368, output_tokens=871
|
|
18:14:24,889 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:24,892 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.46900000050664. input_tokens=2516, output_tokens=683
|
|
18:14:26,899 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:26,904 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 35.59300000034273. input_tokens=2895, output_tokens=824
|
|
18:14:27,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:27,351 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.0. input_tokens=2610, output_tokens=832
|
|
18:14:27,522 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:27,525 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.139999999664724. input_tokens=2570, output_tokens=802
|
|
18:14:27,989 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:27,992 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.73399999924004. input_tokens=4484, output_tokens=1069
|
|
18:14:28,437 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:28,440 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.14099999982864. input_tokens=8012, output_tokens=857
|
|
18:14:28,885 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:28,888 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.51600000075996. input_tokens=7292, output_tokens=858
|
|
18:14:32,349 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:32,352 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 40.985000000335276. input_tokens=2680, output_tokens=839
|
|
18:14:34,849 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:34,852 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.875. input_tokens=2064, output_tokens=442
|
|
18:14:38,573 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:38,575 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.25. input_tokens=2576, output_tokens=664
|
|
18:14:40,215 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:40,218 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.25. input_tokens=2246, output_tokens=773
|
|
18:14:40,285 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:40,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.343999999575317. input_tokens=3615, output_tokens=675
|
|
18:14:42,34 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:42,37 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.514999999664724. input_tokens=2120, output_tokens=609
|
|
18:14:43,131 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:43,133 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.76600000075996. input_tokens=4636, output_tokens=775
|
|
18:14:43,397 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:43,402 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.234000000171363. input_tokens=2422, output_tokens=512
|
|
18:14:43,999 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:44,2 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.53099999949336. input_tokens=2226, output_tokens=730
|
|
18:14:46,859 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:46,860 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.110000000335276. input_tokens=2146, output_tokens=693
|
|
18:14:48,652 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:48,657 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.75. input_tokens=3158, output_tokens=717
|
|
18:14:50,724 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:50,728 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 40.15699999965727. input_tokens=2577, output_tokens=775
|
|
18:14:50,760 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:50,761 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.406000000424683. input_tokens=2951, output_tokens=653
|
|
18:14:50,917 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:50,919 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 39.65599999949336. input_tokens=2595, output_tokens=759
|
|
18:14:50,995 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:50,998 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.561999999918044. input_tokens=2644, output_tokens=685
|
|
18:14:52,816 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:52,820 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.561999999918044. input_tokens=2262, output_tokens=582
|
|
18:14:54,114 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:54,117 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.125. input_tokens=2057, output_tokens=499
|
|
18:14:54,154 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:54,157 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.28100000042468. input_tokens=2475, output_tokens=654
|
|
18:14:54,825 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:54,827 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.468000000342727. input_tokens=2162, output_tokens=429
|
|
18:14:55,376 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:55,379 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.843999999575317. input_tokens=2219, output_tokens=707
|
|
18:14:55,740 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:55,743 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.65599999949336. input_tokens=2976, output_tokens=840
|
|
18:14:57,383 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:57,385 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 35.21900000050664. input_tokens=2565, output_tokens=722
|
|
18:14:57,753 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:57,756 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.875. input_tokens=2443, output_tokens=658
|
|
18:14:58,699 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:58,701 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 38.01500000059605. input_tokens=2462, output_tokens=794
|
|
18:14:59,102 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:59,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.25. input_tokens=2652, output_tokens=835
|
|
18:14:59,530 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:59,533 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.639999999664724. input_tokens=2552, output_tokens=944
|
|
18:14:59,630 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:14:59,633 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.828000000678. input_tokens=3268, output_tokens=715
|
|
18:15:04,674 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:15:04,677 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.093999999575317. input_tokens=2404, output_tokens=807
|
|
18:15:13,30 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:15:13,33 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.811999999918044. input_tokens=9883, output_tokens=807
|
|
18:16:42,883 graphrag.index.cli INFO Logging enabled at E:\uc\Source-Web-23070-DataAnalysis\python\output\indexing-engine.log
|
|
18:16:42,887 graphrag.index.cli INFO Starting pipeline run for: 20250623-181642, dryrun=False
|
|
18:16:42,887 graphrag.index.cli INFO Using default configuration: {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"root_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python",
|
|
"reporting": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"storage": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"cache": {
|
|
"type": "file",
|
|
"base_dir": "cache",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"input": {
|
|
"type": "file",
|
|
"file_type": "text",
|
|
"base_dir": "input",
|
|
"storage_account_blob_url": null,
|
|
"encoding": "utf-8",
|
|
"file_pattern": ".*\\.txt$",
|
|
"file_filter": null,
|
|
"source_column": null,
|
|
"timestamp_column": null,
|
|
"timestamp_format": null,
|
|
"text_column": "text",
|
|
"title_column": null,
|
|
"document_attribute_columns": []
|
|
},
|
|
"embed_graph": {
|
|
"enabled": false,
|
|
"num_walks": 10,
|
|
"walk_length": 40,
|
|
"window_size": 2,
|
|
"iterations": 3,
|
|
"random_seed": 597832,
|
|
"strategy": null
|
|
},
|
|
"embeddings": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_embedding",
|
|
"model": "text-embedding-3-small",
|
|
"max_tokens": 4000,
|
|
"temperature": 0,
|
|
"top_p": 1,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": null,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"batch_size": 16,
|
|
"batch_max_tokens": 8191,
|
|
"target": "required",
|
|
"skip": [],
|
|
"vector_store": null,
|
|
"strategy": null
|
|
},
|
|
"chunks": {
|
|
"size": 1200,
|
|
"overlap": 100,
|
|
"group_by_columns": [
|
|
"id"
|
|
],
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"snapshots": {
|
|
"graphml": true,
|
|
"raw_entities": true,
|
|
"top_level_nodes": true
|
|
},
|
|
"entity_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/entity_extraction.txt",
|
|
"entity_types": [
|
|
"organization",
|
|
"person",
|
|
"geo",
|
|
"event"
|
|
],
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"summarize_descriptions": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/summarize_descriptions.txt",
|
|
"max_length": 500,
|
|
"strategy": null
|
|
},
|
|
"community_reports": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/community_report.txt",
|
|
"max_length": 2000,
|
|
"max_input_length": 8000,
|
|
"strategy": null
|
|
},
|
|
"claim_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"enabled": false,
|
|
"prompt": "prompts/claim_extraction.txt",
|
|
"description": "Any claims or facts that could be relevant to information discovery.",
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"cluster_graph": {
|
|
"max_cluster_size": 10,
|
|
"strategy": null
|
|
},
|
|
"umap": {
|
|
"enabled": false
|
|
},
|
|
"local_search": {
|
|
"text_unit_prop": 0.5,
|
|
"community_prop": 0.1,
|
|
"conversation_history_max_turns": 5,
|
|
"top_k_entities": 10,
|
|
"top_k_relationships": 10,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"llm_max_tokens": 2000
|
|
},
|
|
"global_search": {
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"data_max_tokens": 12000,
|
|
"map_max_tokens": 1000,
|
|
"reduce_max_tokens": 2000,
|
|
"concurrency": 32
|
|
},
|
|
"encoding_model": "cl100k_base",
|
|
"skip_workflows": []
|
|
}
|
|
18:16:42,889 graphrag.index.create_pipeline_config INFO skipping workflows
|
|
18:16:42,889 graphrag.index.run.run INFO Running pipeline
|
|
18:16:42,889 graphrag.index.storage.file_pipeline_storage INFO Creating file storage at E:\uc\Source-Web-23070-DataAnalysis\python\output
|
|
18:16:42,890 graphrag.index.input.load_input INFO loading input from root_dir=input
|
|
18:16:42,890 graphrag.index.input.load_input INFO using file storage for input
|
|
18:16:42,891 graphrag.index.storage.file_pipeline_storage INFO search E:\uc\Source-Web-23070-DataAnalysis\python\input for files matching .*\.txt$
|
|
18:16:42,892 graphrag.index.input.text INFO found text files from input, found [('10_7\u4f5c\u696d\u5206\u4eab.txt', {}), ('\u4f5c\u54c1\u5206\u4eab(09_30).txt', {}), ('\u4f5c\u54c1\u5206\u4eab\uff08\u9ede\u6211\u9032\u5165\u7de8\u8f2f\uff09.txt', {}), ('\u7b2c\u4e00\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1113-14).txt', {}), ('\u7b2c\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e03\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e09\u6b21office hour(1128-1129).txt', {}), ('\u7b2c\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e5d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e8c\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1120-1122).txt', {}), ('\u7b2c\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516b\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {})]
|
|
18:16:42,915 graphrag.index.input.text INFO Found 22 files, loading 22
|
|
18:16:42,917 graphrag.index.workflows.load INFO Workflow Run Order: ['create_base_text_units', 'create_base_extracted_entities', 'create_summarized_entities', 'create_base_entity_graph', 'create_final_entities', 'create_final_nodes', 'create_final_communities', 'create_final_relationships', 'create_final_text_units', 'create_final_community_reports', 'create_base_documents', 'create_final_documents']
|
|
18:16:42,917 graphrag.index.run.run INFO Final # of rows loaded: 22
|
|
18:16:43,63 graphrag.index.run.workflow INFO dependencies for create_base_text_units: []
|
|
18:16:43,68 datashaper.workflow.workflow INFO executing verb orderby
|
|
18:16:43,73 datashaper.workflow.workflow INFO executing verb zip
|
|
18:16:43,77 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
18:16:43,83 datashaper.workflow.workflow INFO executing verb chunk
|
|
18:16:43,497 datashaper.workflow.workflow INFO executing verb select
|
|
18:16:43,502 datashaper.workflow.workflow INFO executing verb unroll
|
|
18:16:43,512 datashaper.workflow.workflow INFO executing verb rename
|
|
18:16:43,521 datashaper.workflow.workflow INFO executing verb genid
|
|
18:16:43,535 datashaper.workflow.workflow INFO executing verb unzip
|
|
18:16:43,546 datashaper.workflow.workflow INFO executing verb copy
|
|
18:16:43,554 datashaper.workflow.workflow INFO executing verb filter
|
|
18:16:43,569 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_text_units.parquet
|
|
18:16:43,731 graphrag.index.run.workflow INFO dependencies for create_base_extracted_entities: ['create_base_text_units']
|
|
18:16:43,731 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:16:43,752 datashaper.workflow.workflow INFO executing verb entity_extract
|
|
18:16:43,767 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:16:43,777 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for gpt-4-turbo-preview: TPM=0, RPM=0
|
|
18:16:43,777 graphrag.index.llm.load_llm INFO create concurrency limiter for gpt-4-turbo-preview: 25
|
|
18:16:44,297 datashaper.workflow.workflow INFO executing verb snapshot
|
|
18:16:44,320 datashaper.workflow.workflow INFO executing verb merge_graphs
|
|
18:16:44,457 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:16:44,463 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_extracted_entities.parquet
|
|
18:16:44,604 graphrag.index.run.workflow INFO dependencies for create_summarized_entities: ['create_base_extracted_entities']
|
|
18:16:44,604 graphrag.utils.storage INFO read table from storage: create_base_extracted_entities.parquet
|
|
18:16:44,626 datashaper.workflow.workflow INFO executing verb summarize_descriptions
|
|
18:16:48,768 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:16:48,772 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.031000000424683. input_tokens=196, output_tokens=73
|
|
18:16:50,117 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:16:50,118 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.358999999240041. input_tokens=230, output_tokens=121
|
|
18:16:50,178 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:16:50,182 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_summarized_entities.parquet
|
|
18:16:50,356 graphrag.index.run.workflow INFO dependencies for create_base_entity_graph: ['create_summarized_entities']
|
|
18:16:50,356 graphrag.utils.storage INFO read table from storage: create_summarized_entities.parquet
|
|
18:16:50,385 datashaper.workflow.workflow INFO executing verb cluster_graph
|
|
18:16:50,843 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:16:50,862 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:16:50,883 datashaper.workflow.workflow INFO executing verb select
|
|
18:16:50,890 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_entity_graph.parquet
|
|
18:16:51,85 graphrag.index.run.workflow INFO dependencies for create_final_entities: ['create_base_entity_graph']
|
|
18:16:51,89 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:16:51,118 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:16:51,300 datashaper.workflow.workflow INFO executing verb rename
|
|
18:16:51,312 datashaper.workflow.workflow INFO executing verb select
|
|
18:16:51,324 datashaper.workflow.workflow INFO executing verb dedupe
|
|
18:16:51,341 datashaper.workflow.workflow INFO executing verb rename
|
|
18:16:51,355 datashaper.workflow.workflow INFO executing verb filter
|
|
18:16:51,388 datashaper.workflow.workflow INFO executing verb text_split
|
|
18:16:51,415 datashaper.workflow.workflow INFO executing verb drop
|
|
18:16:51,433 datashaper.workflow.workflow INFO executing verb merge
|
|
18:16:51,636 datashaper.workflow.workflow INFO executing verb text_embed
|
|
18:16:51,637 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:16:51,649 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for text-embedding-3-small: TPM=0, RPM=0
|
|
18:16:51,649 graphrag.index.llm.load_llm INFO create concurrency limiter for text-embedding-3-small: 25
|
|
18:16:51,713 graphrag.index.verbs.text.embed.strategies.openai INFO embedding 628 inputs via 628 snippets using 40 batches. max_batch_size=16, max_tokens=8191
|
|
18:16:52,905 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:16:53,155 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:16:53,479 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7659999998286366. input_tokens=1060, output_tokens=0
|
|
18:16:53,803 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 2.0779999997466803. input_tokens=798, output_tokens=0
|
|
18:16:53,852 datashaper.workflow.workflow INFO executing verb drop
|
|
18:16:53,869 datashaper.workflow.workflow INFO executing verb filter
|
|
18:16:53,907 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_entities.parquet
|
|
18:16:54,205 graphrag.index.run.workflow INFO dependencies for create_final_nodes: ['create_base_entity_graph']
|
|
18:16:54,211 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:16:54,249 datashaper.workflow.workflow INFO executing verb layout_graph
|
|
18:16:55,103 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:16:55,532 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:16:55,740 datashaper.workflow.workflow INFO executing verb drop
|
|
18:16:55,756 datashaper.workflow.workflow INFO executing verb filter
|
|
18:16:55,846 datashaper.workflow.workflow INFO executing verb select
|
|
18:16:55,870 datashaper.workflow.workflow INFO executing verb snapshot
|
|
18:16:55,897 datashaper.workflow.workflow INFO executing verb rename
|
|
18:16:55,925 datashaper.workflow.workflow INFO executing verb convert
|
|
18:16:55,982 datashaper.workflow.workflow INFO executing verb join
|
|
18:16:56,6 datashaper.workflow.workflow INFO executing verb rename
|
|
18:16:56,9 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_nodes.parquet
|
|
18:16:56,192 graphrag.index.run.workflow INFO dependencies for create_final_communities: ['create_base_entity_graph']
|
|
18:16:56,193 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:16:56,242 datashaper.workflow.workflow INFO executing verb create_final_communities
|
|
18:16:56,850 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_communities.parquet
|
|
18:16:57,49 graphrag.index.run.workflow INFO dependencies for create_final_relationships: ['create_base_entity_graph', 'create_final_nodes']
|
|
18:16:57,50 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:16:57,60 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:16:57,118 datashaper.workflow.workflow INFO executing verb create_final_relationships_pre_embedding
|
|
18:16:57,333 datashaper.workflow.workflow INFO executing verb create_final_relationships_post_embedding
|
|
18:16:57,340 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_relationships.parquet
|
|
18:16:57,537 graphrag.index.run.workflow INFO dependencies for create_final_text_units: ['create_base_text_units', 'create_final_relationships', 'create_final_entities']
|
|
18:16:57,549 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:16:57,554 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:16:57,559 graphrag.utils.storage INFO read table from storage: create_final_entities.parquet
|
|
18:16:57,634 datashaper.workflow.workflow INFO executing verb create_final_text_units_pre_embedding
|
|
18:16:57,710 datashaper.workflow.workflow INFO executing verb select
|
|
18:16:57,713 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_text_units.parquet
|
|
18:16:57,909 graphrag.index.run.workflow INFO dependencies for create_final_community_reports: ['create_final_nodes', 'create_final_relationships']
|
|
18:16:57,910 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:16:57,916 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:16:57,965 datashaper.workflow.workflow INFO executing verb prepare_community_reports_nodes
|
|
18:16:58,27 datashaper.workflow.workflow INFO executing verb prepare_community_reports_edges
|
|
18:16:58,57 datashaper.workflow.workflow INFO executing verb restore_community_hierarchy
|
|
18:16:58,88 datashaper.workflow.workflow INFO executing verb prepare_community_reports
|
|
18:16:58,89 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=3 => 628
|
|
18:16:58,127 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=2 => 628
|
|
18:16:58,221 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=1 => 628
|
|
18:16:58,515 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=0 => 628
|
|
18:16:58,707 datashaper.workflow.workflow INFO executing verb create_community_reports
|
|
18:17:18,138 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:18,141 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.360000000335276. input_tokens=2976, output_tokens=717
|
|
18:17:20,365 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:20,367 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.593999999575317. input_tokens=2492, output_tokens=606
|
|
18:17:34,950 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:34,952 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 35.17200000025332. input_tokens=2611, output_tokens=833
|
|
18:17:51,224 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:51,226 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.093999999575317. input_tokens=2176, output_tokens=531
|
|
18:17:56,589 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:56,591 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.390999999828637. input_tokens=7327, output_tokens=697
|
|
18:17:56,749 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:56,751 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.65599999949336. input_tokens=2197, output_tokens=661
|
|
18:17:57,942 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:57,943 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.859999999403954. input_tokens=5248, output_tokens=689
|
|
18:17:58,13 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:17:58,19 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.828000000678003. input_tokens=2948, output_tokens=718
|
|
18:18:00,777 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:00,779 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.671000000089407. input_tokens=2514, output_tokens=849
|
|
18:18:01,847 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:01,850 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.67200000025332. input_tokens=2764, output_tokens=815
|
|
18:18:01,914 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:01,916 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.734000000171363. input_tokens=2875, output_tokens=772
|
|
18:18:02,396 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:02,399 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.281000000424683. input_tokens=3289, output_tokens=705
|
|
18:18:04,445 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:04,454 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.29700000025332. input_tokens=7464, output_tokens=772
|
|
18:18:04,905 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:04,907 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.734000000171363. input_tokens=3541, output_tokens=915
|
|
18:18:04,974 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:04,976 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.890999999828637. input_tokens=3234, output_tokens=799
|
|
18:18:05,719 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:05,722 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.531000000424683. input_tokens=2547, output_tokens=818
|
|
18:18:06,942 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:06,944 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.828000000678003. input_tokens=3094, output_tokens=856
|
|
18:18:07,22 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:07,39 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.890999999828637. input_tokens=8601, output_tokens=847
|
|
18:18:11,455 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:11,458 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.375. input_tokens=5190, output_tokens=754
|
|
18:18:12,717 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:12,721 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.60900000017136. input_tokens=4890, output_tokens=769
|
|
18:18:13,967 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:13,969 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 38.84400000050664. input_tokens=3407, output_tokens=827
|
|
18:18:16,660 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:16,663 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 41.5. input_tokens=5740, output_tokens=913
|
|
18:18:16,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:18:16,783 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 41.57799999974668. input_tokens=8803, output_tokens=838
|
|
18:18:16,834 datashaper.workflow.workflow INFO executing verb window
|
|
18:18:16,836 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_community_reports.parquet
|
|
18:18:17,73 graphrag.index.run.workflow INFO dependencies for create_base_documents: ['create_final_text_units']
|
|
18:18:17,74 graphrag.utils.storage INFO read table from storage: create_final_text_units.parquet
|
|
18:18:17,136 datashaper.workflow.workflow INFO executing verb unroll
|
|
18:18:17,160 datashaper.workflow.workflow INFO executing verb select
|
|
18:18:17,189 datashaper.workflow.workflow INFO executing verb rename
|
|
18:18:17,212 datashaper.workflow.workflow INFO executing verb join
|
|
18:18:17,239 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
18:18:17,267 datashaper.workflow.workflow INFO executing verb join
|
|
18:18:17,301 datashaper.workflow.workflow INFO executing verb rename
|
|
18:18:17,325 datashaper.workflow.workflow INFO executing verb convert
|
|
18:18:17,378 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_documents.parquet
|
|
18:18:17,588 graphrag.index.run.workflow INFO dependencies for create_final_documents: ['create_base_documents']
|
|
18:18:17,589 graphrag.utils.storage INFO read table from storage: create_base_documents.parquet
|
|
18:18:17,666 datashaper.workflow.workflow INFO executing verb rename
|
|
18:18:17,669 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_documents.parquet
|
|
18:18:17,778 graphrag.index.cli INFO All workflows completed successfully.
|
|
18:21:48,423 graphrag.index.cli INFO Logging enabled at E:\uc\Source-Web-23070-DataAnalysis\python\output\indexing-engine.log
|
|
18:21:48,426 graphrag.index.cli INFO Starting pipeline run for: 20250623-182148, dryrun=False
|
|
18:21:48,427 graphrag.index.cli INFO Using default configuration: {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"root_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python",
|
|
"reporting": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"storage": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"cache": {
|
|
"type": "file",
|
|
"base_dir": "cache",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"input": {
|
|
"type": "file",
|
|
"file_type": "text",
|
|
"base_dir": "input",
|
|
"storage_account_blob_url": null,
|
|
"encoding": "utf-8",
|
|
"file_pattern": ".*\\.txt$",
|
|
"file_filter": null,
|
|
"source_column": null,
|
|
"timestamp_column": null,
|
|
"timestamp_format": null,
|
|
"text_column": "text",
|
|
"title_column": null,
|
|
"document_attribute_columns": []
|
|
},
|
|
"embed_graph": {
|
|
"enabled": true,
|
|
"num_walks": 10,
|
|
"walk_length": 40,
|
|
"window_size": 2,
|
|
"iterations": 3,
|
|
"random_seed": 597832,
|
|
"strategy": null
|
|
},
|
|
"embeddings": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_embedding",
|
|
"model": "text-embedding-3-small",
|
|
"max_tokens": 4000,
|
|
"temperature": 0,
|
|
"top_p": 1,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": null,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"batch_size": 16,
|
|
"batch_max_tokens": 8191,
|
|
"target": "required",
|
|
"skip": [],
|
|
"vector_store": null,
|
|
"strategy": null
|
|
},
|
|
"chunks": {
|
|
"size": 1200,
|
|
"overlap": 100,
|
|
"group_by_columns": [
|
|
"id"
|
|
],
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"snapshots": {
|
|
"graphml": true,
|
|
"raw_entities": true,
|
|
"top_level_nodes": true
|
|
},
|
|
"entity_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/entity_extraction.txt",
|
|
"entity_types": [
|
|
"organization",
|
|
"person",
|
|
"geo",
|
|
"event"
|
|
],
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"summarize_descriptions": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/summarize_descriptions.txt",
|
|
"max_length": 500,
|
|
"strategy": null
|
|
},
|
|
"community_reports": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/community_report.txt",
|
|
"max_length": 2000,
|
|
"max_input_length": 8000,
|
|
"strategy": null
|
|
},
|
|
"claim_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"enabled": false,
|
|
"prompt": "prompts/claim_extraction.txt",
|
|
"description": "Any claims or facts that could be relevant to information discovery.",
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"cluster_graph": {
|
|
"max_cluster_size": 10,
|
|
"strategy": null
|
|
},
|
|
"umap": {
|
|
"enabled": true
|
|
},
|
|
"local_search": {
|
|
"text_unit_prop": 0.5,
|
|
"community_prop": 0.1,
|
|
"conversation_history_max_turns": 5,
|
|
"top_k_entities": 10,
|
|
"top_k_relationships": 10,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"llm_max_tokens": 2000
|
|
},
|
|
"global_search": {
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"data_max_tokens": 12000,
|
|
"map_max_tokens": 1000,
|
|
"reduce_max_tokens": 2000,
|
|
"concurrency": 32
|
|
},
|
|
"encoding_model": "cl100k_base",
|
|
"skip_workflows": []
|
|
}
|
|
18:21:48,428 graphrag.index.create_pipeline_config INFO skipping workflows
|
|
18:21:48,428 graphrag.index.run.run INFO Running pipeline
|
|
18:21:48,429 graphrag.index.storage.file_pipeline_storage INFO Creating file storage at E:\uc\Source-Web-23070-DataAnalysis\python\output
|
|
18:21:48,429 graphrag.index.input.load_input INFO loading input from root_dir=input
|
|
18:21:48,429 graphrag.index.input.load_input INFO using file storage for input
|
|
18:21:48,430 graphrag.index.storage.file_pipeline_storage INFO search E:\uc\Source-Web-23070-DataAnalysis\python\input for files matching .*\.txt$
|
|
18:21:48,431 graphrag.index.input.text INFO found text files from input, found [('10_7\u4f5c\u696d\u5206\u4eab.txt', {}), ('\u4f5c\u54c1\u5206\u4eab(09_30).txt', {}), ('\u4f5c\u54c1\u5206\u4eab\uff08\u9ede\u6211\u9032\u5165\u7de8\u8f2f\uff09.txt', {}), ('\u7b2c\u4e00\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1113-14).txt', {}), ('\u7b2c\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e03\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e09\u6b21office hour(1128-1129).txt', {}), ('\u7b2c\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e5d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e8c\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1120-1122).txt', {}), ('\u7b2c\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516b\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {})]
|
|
18:21:48,450 graphrag.index.input.text INFO Found 22 files, loading 22
|
|
18:21:48,451 graphrag.index.workflows.load INFO Workflow Run Order: ['create_base_text_units', 'create_base_extracted_entities', 'create_summarized_entities', 'create_base_entity_graph', 'create_final_entities', 'create_final_nodes', 'create_final_communities', 'create_final_relationships', 'create_final_text_units', 'create_final_community_reports', 'create_base_documents', 'create_final_documents']
|
|
18:21:48,451 graphrag.index.run.run INFO Final # of rows loaded: 22
|
|
18:21:48,592 graphrag.index.run.workflow INFO dependencies for create_base_text_units: []
|
|
18:21:48,600 datashaper.workflow.workflow INFO executing verb orderby
|
|
18:21:48,606 datashaper.workflow.workflow INFO executing verb zip
|
|
18:21:48,613 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
18:21:48,621 datashaper.workflow.workflow INFO executing verb chunk
|
|
18:21:48,931 datashaper.workflow.workflow INFO executing verb select
|
|
18:21:48,936 datashaper.workflow.workflow INFO executing verb unroll
|
|
18:21:48,943 datashaper.workflow.workflow INFO executing verb rename
|
|
18:21:48,948 datashaper.workflow.workflow INFO executing verb genid
|
|
18:21:48,959 datashaper.workflow.workflow INFO executing verb unzip
|
|
18:21:48,966 datashaper.workflow.workflow INFO executing verb copy
|
|
18:21:48,973 datashaper.workflow.workflow INFO executing verb filter
|
|
18:21:48,986 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_text_units.parquet
|
|
18:21:49,153 graphrag.index.run.workflow INFO dependencies for create_base_extracted_entities: ['create_base_text_units']
|
|
18:21:49,153 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:21:49,180 datashaper.workflow.workflow INFO executing verb entity_extract
|
|
18:21:49,200 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:21:49,216 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for gpt-4-turbo-preview: TPM=0, RPM=0
|
|
18:21:49,216 graphrag.index.llm.load_llm INFO create concurrency limiter for gpt-4-turbo-preview: 25
|
|
18:21:49,716 datashaper.workflow.workflow INFO executing verb snapshot
|
|
18:21:49,750 datashaper.workflow.workflow INFO executing verb merge_graphs
|
|
18:21:49,913 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:21:49,917 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_extracted_entities.parquet
|
|
18:21:50,54 graphrag.index.run.workflow INFO dependencies for create_summarized_entities: ['create_base_extracted_entities']
|
|
18:21:50,55 graphrag.utils.storage INFO read table from storage: create_base_extracted_entities.parquet
|
|
18:21:50,75 datashaper.workflow.workflow INFO executing verb summarize_descriptions
|
|
18:21:50,567 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:21:50,571 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_summarized_entities.parquet
|
|
18:21:50,722 graphrag.index.run.workflow INFO dependencies for create_base_entity_graph: ['create_summarized_entities']
|
|
18:21:50,722 graphrag.utils.storage INFO read table from storage: create_summarized_entities.parquet
|
|
18:21:50,745 datashaper.workflow.workflow INFO executing verb cluster_graph
|
|
18:21:51,235 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:21:51,252 datashaper.workflow.workflow INFO executing verb embed_graph
|
|
18:21:51,446 root INFO Starting preprocessing of transition probabilities on graph with 357 nodes and 547 edges
|
|
18:21:51,446 root INFO Starting at time 1750674111.4465835
|
|
18:21:51,446 root INFO Beginning preprocessing of transition probabilities for 357 vertices
|
|
18:21:51,446 root INFO Completed 1 / 357 vertices
|
|
18:21:51,447 root INFO Completed 36 / 357 vertices
|
|
18:21:51,452 root INFO Completed 71 / 357 vertices
|
|
18:21:51,453 root INFO Completed 106 / 357 vertices
|
|
18:21:51,453 root INFO Completed 141 / 357 vertices
|
|
18:21:51,453 root INFO Completed 176 / 357 vertices
|
|
18:21:51,454 root INFO Completed 211 / 357 vertices
|
|
18:21:51,457 root INFO Completed 246 / 357 vertices
|
|
18:21:51,458 root INFO Completed 281 / 357 vertices
|
|
18:21:51,458 root INFO Completed 316 / 357 vertices
|
|
18:21:51,458 root INFO Completed 351 / 357 vertices
|
|
18:21:51,458 root INFO Completed preprocessing of transition probabilities for vertices
|
|
18:21:51,458 root INFO Beginning preprocessing of transition probabilities for 547 edges
|
|
18:21:51,458 root INFO Completed 1 / 547 edges
|
|
18:21:51,461 root INFO Completed 55 / 547 edges
|
|
18:21:51,464 root INFO Completed 109 / 547 edges
|
|
18:21:51,467 root INFO Completed 163 / 547 edges
|
|
18:21:51,471 root INFO Completed 217 / 547 edges
|
|
18:21:51,474 root INFO Completed 271 / 547 edges
|
|
18:21:51,476 root INFO Completed 325 / 547 edges
|
|
18:21:51,479 root INFO Completed 379 / 547 edges
|
|
18:21:51,479 root INFO Completed 433 / 547 edges
|
|
18:21:51,480 root INFO Completed 487 / 547 edges
|
|
18:21:51,481 root INFO Completed 541 / 547 edges
|
|
18:21:51,482 root INFO Completed preprocessing of transition probabilities for edges
|
|
18:21:51,482 root INFO Simulating walks on graph at time 1750674111.482173
|
|
18:21:51,482 root INFO Walk iteration: 1/10
|
|
18:21:51,505 root INFO Walk iteration: 2/10
|
|
18:21:51,526 root INFO Walk iteration: 3/10
|
|
18:21:51,561 root INFO Walk iteration: 4/10
|
|
18:21:51,589 root INFO Walk iteration: 5/10
|
|
18:21:51,615 root INFO Walk iteration: 6/10
|
|
18:21:51,647 root INFO Walk iteration: 7/10
|
|
18:21:51,672 root INFO Walk iteration: 8/10
|
|
18:21:51,696 root INFO Walk iteration: 9/10
|
|
18:21:51,726 root INFO Walk iteration: 10/10
|
|
18:21:51,746 root INFO Learning embeddings at time 1750674111.7469363
|
|
18:21:51,751 gensim.models.word2vec INFO collecting all words and their counts
|
|
18:21:51,751 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
18:21:51,755 gensim.models.word2vec INFO collected 357 word types from a corpus of 67440 raw words and 3570 sentences
|
|
18:21:51,755 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
18:21:51,756 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 357 unique words (100.00% of original 357, drops 0)', 'datetime': '2025-06-23T18:21:51.756430', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:51,756 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 67440 word corpus (100.00% of original 67440, drops 0)', 'datetime': '2025-06-23T18:21:51.756430', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:51,757 gensim.models.word2vec INFO deleting the raw counts dictionary of 357 items
|
|
18:21:51,757 gensim.models.word2vec INFO sample=0.001 downsamples 94 most-common words
|
|
18:21:51,757 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 40193.71660964316 word corpus (59.6%% of prior 67440)', 'datetime': '2025-06-23T18:21:51.757893', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:51,760 gensim.models.word2vec INFO estimated required memory for 357 words and 1536 dimensions: 4565316 bytes
|
|
18:21:51,760 gensim.models.word2vec INFO resetting layer weights
|
|
18:21:51,762 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-06-23T18:21:51.762399', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
18:21:51,763 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 357 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-06-23T18:21:51.763120', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:51,863 gensim.models.word2vec INFO EPOCH 0: training on 67440 raw words (40344 effective words) took 0.1s, 424731 effective words/s
|
|
18:21:51,963 gensim.models.word2vec INFO EPOCH 1: training on 67440 raw words (40259 effective words) took 0.1s, 422324 effective words/s
|
|
18:21:52,61 gensim.models.word2vec INFO EPOCH 2: training on 67440 raw words (40011 effective words) took 0.1s, 432005 effective words/s
|
|
18:21:52,61 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 202320 raw words (120614 effective words) took 0.3s, 404331 effective words/s', 'datetime': '2025-06-23T18:21:52.061205', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:52,61 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=357, vector_size=1536, alpha=0.025>', 'datetime': '2025-06-23T18:21:52.061205', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
18:21:52,61 root INFO Completed. Ending time is 1750674112.0612054 Elapsed time is -0.6146218776702881
|
|
18:21:52,137 root INFO Starting preprocessing of transition probabilities on graph with 357 nodes and 547 edges
|
|
18:21:52,137 root INFO Starting at time 1750674112.1371443
|
|
18:21:52,137 root INFO Beginning preprocessing of transition probabilities for 357 vertices
|
|
18:21:52,137 root INFO Completed 1 / 357 vertices
|
|
18:21:52,137 root INFO Completed 36 / 357 vertices
|
|
18:21:52,137 root INFO Completed 71 / 357 vertices
|
|
18:21:52,138 root INFO Completed 106 / 357 vertices
|
|
18:21:52,138 root INFO Completed 141 / 357 vertices
|
|
18:21:52,139 root INFO Completed 176 / 357 vertices
|
|
18:21:52,139 root INFO Completed 211 / 357 vertices
|
|
18:21:52,140 root INFO Completed 246 / 357 vertices
|
|
18:21:52,140 root INFO Completed 281 / 357 vertices
|
|
18:21:52,140 root INFO Completed 316 / 357 vertices
|
|
18:21:52,141 root INFO Completed 351 / 357 vertices
|
|
18:21:52,141 root INFO Completed preprocessing of transition probabilities for vertices
|
|
18:21:52,141 root INFO Beginning preprocessing of transition probabilities for 547 edges
|
|
18:21:52,141 root INFO Completed 1 / 547 edges
|
|
18:21:52,149 root INFO Completed 55 / 547 edges
|
|
18:21:52,153 root INFO Completed 109 / 547 edges
|
|
18:21:52,158 root INFO Completed 163 / 547 edges
|
|
18:21:52,169 root INFO Completed 217 / 547 edges
|
|
18:21:52,175 root INFO Completed 271 / 547 edges
|
|
18:21:52,177 root INFO Completed 325 / 547 edges
|
|
18:21:52,181 root INFO Completed 379 / 547 edges
|
|
18:21:52,183 root INFO Completed 433 / 547 edges
|
|
18:21:52,186 root INFO Completed 487 / 547 edges
|
|
18:21:52,189 root INFO Completed 541 / 547 edges
|
|
18:21:52,190 root INFO Completed preprocessing of transition probabilities for edges
|
|
18:21:52,190 root INFO Simulating walks on graph at time 1750674112.1904259
|
|
18:21:52,191 root INFO Walk iteration: 1/10
|
|
18:21:52,234 root INFO Walk iteration: 2/10
|
|
18:21:52,278 root INFO Walk iteration: 3/10
|
|
18:21:52,319 root INFO Walk iteration: 4/10
|
|
18:21:52,357 root INFO Walk iteration: 5/10
|
|
18:21:52,386 root INFO Walk iteration: 6/10
|
|
18:21:52,407 root INFO Walk iteration: 7/10
|
|
18:21:52,431 root INFO Walk iteration: 8/10
|
|
18:21:52,455 root INFO Walk iteration: 9/10
|
|
18:21:52,476 root INFO Walk iteration: 10/10
|
|
18:21:52,498 root INFO Learning embeddings at time 1750674112.4982193
|
|
18:21:52,502 gensim.models.word2vec INFO collecting all words and their counts
|
|
18:21:52,502 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
18:21:52,507 gensim.models.word2vec INFO collected 357 word types from a corpus of 67440 raw words and 3570 sentences
|
|
18:21:52,507 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
18:21:52,507 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 357 unique words (100.00% of original 357, drops 0)', 'datetime': '2025-06-23T18:21:52.507917', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:52,507 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 67440 word corpus (100.00% of original 67440, drops 0)', 'datetime': '2025-06-23T18:21:52.507917', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:52,509 gensim.models.word2vec INFO deleting the raw counts dictionary of 357 items
|
|
18:21:52,509 gensim.models.word2vec INFO sample=0.001 downsamples 94 most-common words
|
|
18:21:52,509 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 40193.71660964316 word corpus (59.6%% of prior 67440)', 'datetime': '2025-06-23T18:21:52.509410', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:52,511 gensim.models.word2vec INFO estimated required memory for 357 words and 1536 dimensions: 4565316 bytes
|
|
18:21:52,511 gensim.models.word2vec INFO resetting layer weights
|
|
18:21:52,513 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-06-23T18:21:52.513890', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
18:21:52,513 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 357 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-06-23T18:21:52.513890', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:52,610 gensim.models.word2vec INFO EPOCH 0: training on 67440 raw words (40188 effective words) took 0.1s, 438228 effective words/s
|
|
18:21:52,713 gensim.models.word2vec INFO EPOCH 1: training on 67440 raw words (40312 effective words) took 0.1s, 414331 effective words/s
|
|
18:21:52,812 gensim.models.word2vec INFO EPOCH 2: training on 67440 raw words (40163 effective words) took 0.1s, 424847 effective words/s
|
|
18:21:52,812 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 202320 raw words (120663 effective words) took 0.3s, 403754 effective words/s', 'datetime': '2025-06-23T18:21:52.812705', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:52,812 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=357, vector_size=1536, alpha=0.025>', 'datetime': '2025-06-23T18:21:52.812705', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
18:21:52,813 root INFO Completed. Ending time is 1750674112.8134527 Elapsed time is -0.6763083934783936
|
|
18:21:52,914 root INFO Starting preprocessing of transition probabilities on graph with 357 nodes and 547 edges
|
|
18:21:52,914 root INFO Starting at time 1750674112.9142623
|
|
18:21:52,914 root INFO Beginning preprocessing of transition probabilities for 357 vertices
|
|
18:21:52,914 root INFO Completed 1 / 357 vertices
|
|
18:21:52,915 root INFO Completed 36 / 357 vertices
|
|
18:21:52,916 root INFO Completed 71 / 357 vertices
|
|
18:21:52,917 root INFO Completed 106 / 357 vertices
|
|
18:21:52,917 root INFO Completed 141 / 357 vertices
|
|
18:21:52,918 root INFO Completed 176 / 357 vertices
|
|
18:21:52,918 root INFO Completed 211 / 357 vertices
|
|
18:21:52,919 root INFO Completed 246 / 357 vertices
|
|
18:21:52,920 root INFO Completed 281 / 357 vertices
|
|
18:21:52,920 root INFO Completed 316 / 357 vertices
|
|
18:21:52,920 root INFO Completed 351 / 357 vertices
|
|
18:21:52,920 root INFO Completed preprocessing of transition probabilities for vertices
|
|
18:21:52,920 root INFO Beginning preprocessing of transition probabilities for 547 edges
|
|
18:21:52,920 root INFO Completed 1 / 547 edges
|
|
18:21:52,926 root INFO Completed 55 / 547 edges
|
|
18:21:52,928 root INFO Completed 109 / 547 edges
|
|
18:21:52,934 root INFO Completed 163 / 547 edges
|
|
18:21:52,940 root INFO Completed 217 / 547 edges
|
|
18:21:52,943 root INFO Completed 271 / 547 edges
|
|
18:21:52,945 root INFO Completed 325 / 547 edges
|
|
18:21:52,948 root INFO Completed 379 / 547 edges
|
|
18:21:52,950 root INFO Completed 433 / 547 edges
|
|
18:21:52,950 root INFO Completed 487 / 547 edges
|
|
18:21:52,952 root INFO Completed 541 / 547 edges
|
|
18:21:52,952 root INFO Completed preprocessing of transition probabilities for edges
|
|
18:21:52,952 root INFO Simulating walks on graph at time 1750674112.952465
|
|
18:21:52,952 root INFO Walk iteration: 1/10
|
|
18:21:52,983 root INFO Walk iteration: 2/10
|
|
18:21:53,8 root INFO Walk iteration: 3/10
|
|
18:21:53,27 root INFO Walk iteration: 4/10
|
|
18:21:53,47 root INFO Walk iteration: 5/10
|
|
18:21:53,77 root INFO Walk iteration: 6/10
|
|
18:21:53,104 root INFO Walk iteration: 7/10
|
|
18:21:53,128 root INFO Walk iteration: 8/10
|
|
18:21:53,148 root INFO Walk iteration: 9/10
|
|
18:21:53,168 root INFO Walk iteration: 10/10
|
|
18:21:53,187 root INFO Learning embeddings at time 1750674113.187364
|
|
18:21:53,191 gensim.models.word2vec INFO collecting all words and their counts
|
|
18:21:53,191 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
18:21:53,196 gensim.models.word2vec INFO collected 357 word types from a corpus of 67440 raw words and 3570 sentences
|
|
18:21:53,196 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
18:21:53,197 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 357 unique words (100.00% of original 357, drops 0)', 'datetime': '2025-06-23T18:21:53.197070', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,197 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 67440 word corpus (100.00% of original 67440, drops 0)', 'datetime': '2025-06-23T18:21:53.197070', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,198 gensim.models.word2vec INFO deleting the raw counts dictionary of 357 items
|
|
18:21:53,199 gensim.models.word2vec INFO sample=0.001 downsamples 94 most-common words
|
|
18:21:53,199 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 40193.71660964316 word corpus (59.6%% of prior 67440)', 'datetime': '2025-06-23T18:21:53.199069', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,200 gensim.models.word2vec INFO estimated required memory for 357 words and 1536 dimensions: 4565316 bytes
|
|
18:21:53,200 gensim.models.word2vec INFO resetting layer weights
|
|
18:21:53,202 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-06-23T18:21:53.202802', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
18:21:53,202 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 357 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-06-23T18:21:53.202802', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:53,313 gensim.models.word2vec INFO EPOCH 0: training on 67440 raw words (40134 effective words) took 0.1s, 380669 effective words/s
|
|
18:21:53,424 gensim.models.word2vec INFO EPOCH 1: training on 67440 raw words (40333 effective words) took 0.1s, 383039 effective words/s
|
|
18:21:53,539 gensim.models.word2vec INFO EPOCH 2: training on 67440 raw words (40135 effective words) took 0.1s, 366363 effective words/s
|
|
18:21:53,539 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 202320 raw words (120602 effective words) took 0.3s, 358336 effective words/s', 'datetime': '2025-06-23T18:21:53.539207', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:53,539 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=357, vector_size=1536, alpha=0.025>', 'datetime': '2025-06-23T18:21:53.539207', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
18:21:53,539 root INFO Completed. Ending time is 1750674113.5399544 Elapsed time is -0.6256921291351318
|
|
18:21:53,606 root INFO Starting preprocessing of transition probabilities on graph with 357 nodes and 547 edges
|
|
18:21:53,606 root INFO Starting at time 1750674113.6066718
|
|
18:21:53,606 root INFO Beginning preprocessing of transition probabilities for 357 vertices
|
|
18:21:53,606 root INFO Completed 1 / 357 vertices
|
|
18:21:53,607 root INFO Completed 36 / 357 vertices
|
|
18:21:53,607 root INFO Completed 71 / 357 vertices
|
|
18:21:53,608 root INFO Completed 106 / 357 vertices
|
|
18:21:53,608 root INFO Completed 141 / 357 vertices
|
|
18:21:53,608 root INFO Completed 176 / 357 vertices
|
|
18:21:53,608 root INFO Completed 211 / 357 vertices
|
|
18:21:53,608 root INFO Completed 246 / 357 vertices
|
|
18:21:53,608 root INFO Completed 281 / 357 vertices
|
|
18:21:53,608 root INFO Completed 316 / 357 vertices
|
|
18:21:53,609 root INFO Completed 351 / 357 vertices
|
|
18:21:53,609 root INFO Completed preprocessing of transition probabilities for vertices
|
|
18:21:53,609 root INFO Beginning preprocessing of transition probabilities for 547 edges
|
|
18:21:53,609 root INFO Completed 1 / 547 edges
|
|
18:21:53,612 root INFO Completed 55 / 547 edges
|
|
18:21:53,614 root INFO Completed 109 / 547 edges
|
|
18:21:53,618 root INFO Completed 163 / 547 edges
|
|
18:21:53,626 root INFO Completed 217 / 547 edges
|
|
18:21:53,631 root INFO Completed 271 / 547 edges
|
|
18:21:53,635 root INFO Completed 325 / 547 edges
|
|
18:21:53,637 root INFO Completed 379 / 547 edges
|
|
18:21:53,638 root INFO Completed 433 / 547 edges
|
|
18:21:53,639 root INFO Completed 487 / 547 edges
|
|
18:21:53,640 root INFO Completed 541 / 547 edges
|
|
18:21:53,640 root INFO Completed preprocessing of transition probabilities for edges
|
|
18:21:53,640 root INFO Simulating walks on graph at time 1750674113.6402714
|
|
18:21:53,641 root INFO Walk iteration: 1/10
|
|
18:21:53,660 root INFO Walk iteration: 2/10
|
|
18:21:53,679 root INFO Walk iteration: 3/10
|
|
18:21:53,699 root INFO Walk iteration: 4/10
|
|
18:21:53,728 root INFO Walk iteration: 5/10
|
|
18:21:53,749 root INFO Walk iteration: 6/10
|
|
18:21:53,770 root INFO Walk iteration: 7/10
|
|
18:21:53,790 root INFO Walk iteration: 8/10
|
|
18:21:53,813 root INFO Walk iteration: 9/10
|
|
18:21:53,833 root INFO Walk iteration: 10/10
|
|
18:21:53,859 root INFO Learning embeddings at time 1750674113.8590417
|
|
18:21:53,863 gensim.models.word2vec INFO collecting all words and their counts
|
|
18:21:53,864 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
18:21:53,870 gensim.models.word2vec INFO collected 357 word types from a corpus of 67440 raw words and 3570 sentences
|
|
18:21:53,870 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
18:21:53,870 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 357 unique words (100.00% of original 357, drops 0)', 'datetime': '2025-06-23T18:21:53.870943', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,870 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 67440 word corpus (100.00% of original 67440, drops 0)', 'datetime': '2025-06-23T18:21:53.870943', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,872 gensim.models.word2vec INFO deleting the raw counts dictionary of 357 items
|
|
18:21:53,872 gensim.models.word2vec INFO sample=0.001 downsamples 94 most-common words
|
|
18:21:53,872 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 40193.71660964316 word corpus (59.6%% of prior 67440)', 'datetime': '2025-06-23T18:21:53.872438', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
18:21:53,874 gensim.models.word2vec INFO estimated required memory for 357 words and 1536 dimensions: 4565316 bytes
|
|
18:21:53,874 gensim.models.word2vec INFO resetting layer weights
|
|
18:21:53,876 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-06-23T18:21:53.876140', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
18:21:53,876 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 357 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-06-23T18:21:53.876140', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:53,979 gensim.models.word2vec INFO EPOCH 0: training on 67440 raw words (40247 effective words) took 0.1s, 411812 effective words/s
|
|
18:21:54,88 gensim.models.word2vec INFO EPOCH 1: training on 67440 raw words (40152 effective words) took 0.1s, 389733 effective words/s
|
|
18:21:54,193 gensim.models.word2vec INFO EPOCH 2: training on 67440 raw words (40212 effective words) took 0.1s, 400053 effective words/s
|
|
18:21:54,193 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 202320 raw words (120611 effective words) took 0.3s, 380184 effective words/s', 'datetime': '2025-06-23T18:21:54.193863', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
18:21:54,193 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=357, vector_size=1536, alpha=0.025>', 'datetime': '2025-06-23T18:21:54.193863', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
18:21:54,193 root INFO Completed. Ending time is 1750674114.193863 Elapsed time is -0.587191104888916
|
|
18:21:54,221 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
18:21:54,240 datashaper.workflow.workflow INFO executing verb select
|
|
18:21:54,291 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_entity_graph.parquet
|
|
18:21:55,136 graphrag.index.run.workflow INFO dependencies for create_final_entities: ['create_base_entity_graph']
|
|
18:21:55,136 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:21:55,300 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:21:55,490 datashaper.workflow.workflow INFO executing verb rename
|
|
18:21:55,503 datashaper.workflow.workflow INFO executing verb select
|
|
18:21:55,515 datashaper.workflow.workflow INFO executing verb dedupe
|
|
18:21:55,531 datashaper.workflow.workflow INFO executing verb rename
|
|
18:21:55,544 datashaper.workflow.workflow INFO executing verb filter
|
|
18:21:55,584 datashaper.workflow.workflow INFO executing verb text_split
|
|
18:21:55,614 datashaper.workflow.workflow INFO executing verb drop
|
|
18:21:55,629 datashaper.workflow.workflow INFO executing verb merge
|
|
18:21:55,776 datashaper.workflow.workflow INFO executing verb text_embed
|
|
18:21:55,777 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
18:21:55,786 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for text-embedding-3-small: TPM=0, RPM=0
|
|
18:21:55,786 graphrag.index.llm.load_llm INFO create concurrency limiter for text-embedding-3-small: 25
|
|
18:21:55,863 graphrag.index.verbs.text.embed.strategies.openai INFO embedding 628 inputs via 628 snippets using 40 batches. max_batch_size=16, max_tokens=8191
|
|
18:21:57,257 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
18:21:57,822 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.9530000006780028. input_tokens=1138, output_tokens=0
|
|
18:21:57,879 datashaper.workflow.workflow INFO executing verb drop
|
|
18:21:57,898 datashaper.workflow.workflow INFO executing verb filter
|
|
18:21:57,930 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_entities.parquet
|
|
18:21:58,274 graphrag.index.run.workflow INFO dependencies for create_final_nodes: ['create_base_entity_graph']
|
|
18:21:58,274 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:21:58,461 datashaper.workflow.workflow INFO executing verb layout_graph
|
|
18:22:11,0 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:22:11,450 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
18:22:11,684 datashaper.workflow.workflow INFO executing verb drop
|
|
18:22:11,700 datashaper.workflow.workflow INFO executing verb filter
|
|
18:22:11,768 datashaper.workflow.workflow INFO executing verb select
|
|
18:22:11,784 datashaper.workflow.workflow INFO executing verb snapshot
|
|
18:22:11,802 datashaper.workflow.workflow INFO executing verb rename
|
|
18:22:11,818 datashaper.workflow.workflow INFO executing verb join
|
|
18:22:11,842 datashaper.workflow.workflow INFO executing verb convert
|
|
18:22:11,912 datashaper.workflow.workflow INFO executing verb rename
|
|
18:22:11,917 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_nodes.parquet
|
|
18:22:12,354 graphrag.index.run.workflow INFO dependencies for create_final_communities: ['create_base_entity_graph']
|
|
18:22:12,355 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:22:12,511 datashaper.workflow.workflow INFO executing verb create_final_communities
|
|
18:22:13,101 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_communities.parquet
|
|
18:22:13,334 graphrag.index.run.workflow INFO dependencies for create_final_relationships: ['create_final_nodes', 'create_base_entity_graph']
|
|
18:22:13,335 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:22:13,474 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
18:22:13,640 datashaper.workflow.workflow INFO executing verb create_final_relationships_pre_embedding
|
|
18:22:13,959 datashaper.workflow.workflow INFO executing verb create_final_relationships_post_embedding
|
|
18:22:13,971 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_relationships.parquet
|
|
18:22:14,223 graphrag.index.run.workflow INFO dependencies for create_final_text_units: ['create_final_relationships', 'create_base_text_units', 'create_final_entities']
|
|
18:22:14,237 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:22:14,243 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
18:22:14,251 graphrag.utils.storage INFO read table from storage: create_final_entities.parquet
|
|
18:22:14,381 datashaper.workflow.workflow INFO executing verb create_final_text_units_pre_embedding
|
|
18:22:14,462 datashaper.workflow.workflow INFO executing verb select
|
|
18:22:14,466 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_text_units.parquet
|
|
18:22:14,706 graphrag.index.run.workflow INFO dependencies for create_final_community_reports: ['create_final_relationships', 'create_final_nodes']
|
|
18:22:14,707 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
18:22:14,713 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
18:22:14,848 datashaper.workflow.workflow INFO executing verb prepare_community_reports_nodes
|
|
18:22:14,928 datashaper.workflow.workflow INFO executing verb prepare_community_reports_edges
|
|
18:22:14,994 datashaper.workflow.workflow INFO executing verb restore_community_hierarchy
|
|
18:22:15,60 datashaper.workflow.workflow INFO executing verb prepare_community_reports
|
|
18:22:15,62 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=3 => 628
|
|
18:22:15,159 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=2 => 628
|
|
18:22:15,283 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=1 => 628
|
|
18:22:15,526 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=0 => 628
|
|
18:22:15,761 datashaper.workflow.workflow INFO executing verb create_community_reports
|
|
18:22:36,805 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:22:36,809 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.843999999575317. input_tokens=2611, output_tokens=687
|
|
18:22:45,610 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:22:45,613 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.625. input_tokens=2570, output_tokens=824
|
|
18:23:09,782 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:23:09,783 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.968000000342727. input_tokens=5818, output_tokens=780
|
|
18:23:10,143 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:23:10,145 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.34400000050664. input_tokens=5248, output_tokens=816
|
|
18:23:20,677 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
18:23:20,680 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.875. input_tokens=2547, output_tokens=795
|
|
18:23:20,740 datashaper.workflow.workflow INFO executing verb window
|
|
18:23:20,743 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_community_reports.parquet
|
|
18:23:20,974 graphrag.index.run.workflow INFO dependencies for create_base_documents: ['create_final_text_units']
|
|
18:23:20,975 graphrag.utils.storage INFO read table from storage: create_final_text_units.parquet
|
|
18:23:21,26 datashaper.workflow.workflow INFO executing verb unroll
|
|
18:23:21,50 datashaper.workflow.workflow INFO executing verb select
|
|
18:23:21,79 datashaper.workflow.workflow INFO executing verb rename
|
|
18:23:21,106 datashaper.workflow.workflow INFO executing verb join
|
|
18:23:21,135 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
18:23:21,161 datashaper.workflow.workflow INFO executing verb join
|
|
18:23:21,189 datashaper.workflow.workflow INFO executing verb rename
|
|
18:23:21,214 datashaper.workflow.workflow INFO executing verb convert
|
|
18:23:21,357 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_documents.parquet
|
|
18:23:21,577 graphrag.index.run.workflow INFO dependencies for create_final_documents: ['create_base_documents']
|
|
18:23:21,577 graphrag.utils.storage INFO read table from storage: create_base_documents.parquet
|
|
18:23:21,632 datashaper.workflow.workflow INFO executing verb rename
|
|
18:23:21,633 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_documents.parquet
|
|
18:23:21,732 graphrag.index.cli INFO All workflows completed successfully.
|
|
15:29:10,234 graphrag.index.cli INFO Logging enabled at E:\uc\Source-Web-23070-DataAnalysis\python\output\indexing-engine.log
|
|
15:29:10,238 graphrag.index.cli INFO Starting pipeline run for: 20250703-152910, dryrun=False
|
|
15:29:10,240 graphrag.index.cli INFO Using default configuration: {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"root_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python",
|
|
"reporting": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"storage": {
|
|
"type": "file",
|
|
"base_dir": "E:\\uc\\Source-Web-23070-DataAnalysis\\python\\output",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"cache": {
|
|
"type": "file",
|
|
"base_dir": "cache",
|
|
"storage_account_blob_url": null
|
|
},
|
|
"input": {
|
|
"type": "file",
|
|
"file_type": "text",
|
|
"base_dir": "input",
|
|
"storage_account_blob_url": null,
|
|
"encoding": "utf-8",
|
|
"file_pattern": ".*\\.txt$",
|
|
"file_filter": null,
|
|
"source_column": null,
|
|
"timestamp_column": null,
|
|
"timestamp_format": null,
|
|
"text_column": "text",
|
|
"title_column": null,
|
|
"document_attribute_columns": []
|
|
},
|
|
"embed_graph": {
|
|
"enabled": true,
|
|
"num_walks": 10,
|
|
"walk_length": 40,
|
|
"window_size": 2,
|
|
"iterations": 3,
|
|
"random_seed": 597832,
|
|
"strategy": null
|
|
},
|
|
"embeddings": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_embedding",
|
|
"model": "text-embedding-3-small",
|
|
"max_tokens": 4000,
|
|
"temperature": 0,
|
|
"top_p": 1,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": null,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"batch_size": 16,
|
|
"batch_max_tokens": 8191,
|
|
"target": "required",
|
|
"skip": [],
|
|
"vector_store": null,
|
|
"strategy": null
|
|
},
|
|
"chunks": {
|
|
"size": 1200,
|
|
"overlap": 100,
|
|
"group_by_columns": [
|
|
"id"
|
|
],
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"snapshots": {
|
|
"graphml": true,
|
|
"raw_entities": true,
|
|
"top_level_nodes": true
|
|
},
|
|
"entity_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/entity_extraction.txt",
|
|
"entity_types": [
|
|
"organization",
|
|
"person",
|
|
"geo",
|
|
"event"
|
|
],
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"summarize_descriptions": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/summarize_descriptions.txt",
|
|
"max_length": 500,
|
|
"strategy": null
|
|
},
|
|
"community_reports": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"prompt": "prompts/community_report.txt",
|
|
"max_length": 2000,
|
|
"max_input_length": 8000,
|
|
"strategy": null
|
|
},
|
|
"claim_extraction": {
|
|
"llm": {
|
|
"api_key": "==== REDACTED ====",
|
|
"type": "openai_chat",
|
|
"model": "gpt-4-turbo-preview",
|
|
"max_tokens": 4000,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"request_timeout": 180.0,
|
|
"api_base": null,
|
|
"api_version": null,
|
|
"proxy": null,
|
|
"cognitive_services_endpoint": null,
|
|
"deployment_name": null,
|
|
"model_supports_json": true,
|
|
"tokens_per_minute": 0,
|
|
"requests_per_minute": 0,
|
|
"max_retries": 10,
|
|
"max_retry_wait": 10.0,
|
|
"sleep_on_rate_limit_recommendation": true,
|
|
"concurrent_requests": 25
|
|
},
|
|
"parallelization": {
|
|
"stagger": 0.3,
|
|
"num_threads": 50
|
|
},
|
|
"async_mode": "threaded",
|
|
"enabled": false,
|
|
"prompt": "prompts/claim_extraction.txt",
|
|
"description": "Any claims or facts that could be relevant to information discovery.",
|
|
"max_gleanings": 1,
|
|
"strategy": null,
|
|
"encoding_model": null
|
|
},
|
|
"cluster_graph": {
|
|
"max_cluster_size": 10,
|
|
"strategy": null
|
|
},
|
|
"umap": {
|
|
"enabled": true
|
|
},
|
|
"local_search": {
|
|
"text_unit_prop": 0.5,
|
|
"community_prop": 0.1,
|
|
"conversation_history_max_turns": 5,
|
|
"top_k_entities": 10,
|
|
"top_k_relationships": 10,
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"llm_max_tokens": 2000
|
|
},
|
|
"global_search": {
|
|
"temperature": 0.0,
|
|
"top_p": 1.0,
|
|
"n": 1,
|
|
"max_tokens": 12000,
|
|
"data_max_tokens": 12000,
|
|
"map_max_tokens": 1000,
|
|
"reduce_max_tokens": 2000,
|
|
"concurrency": 32
|
|
},
|
|
"encoding_model": "cl100k_base",
|
|
"skip_workflows": []
|
|
}
|
|
15:29:10,243 graphrag.index.create_pipeline_config INFO skipping workflows
|
|
15:29:10,243 graphrag.index.run.run INFO Running pipeline
|
|
15:29:10,243 graphrag.index.storage.file_pipeline_storage INFO Creating file storage at E:\uc\Source-Web-23070-DataAnalysis\python\output
|
|
15:29:10,243 graphrag.index.input.load_input INFO loading input from root_dir=input
|
|
15:29:10,243 graphrag.index.input.load_input INFO using file storage for input
|
|
15:29:10,245 graphrag.index.storage.file_pipeline_storage INFO search E:\uc\Source-Web-23070-DataAnalysis\python\input for files matching .*\.txt$
|
|
15:29:10,246 graphrag.index.input.text INFO found text files from input, found [('10_7\u4f5c\u696d\u5206\u4eab.txt', {}), ('\u4f5c\u54c1\u5206\u4eab(09_30).txt', {}), ('\u4f5c\u54c1\u5206\u4eab\uff08\u9ede\u6211\u9032\u5165\u7de8\u8f2f\uff09.txt', {}), ('\u7b2c\u4e00\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1113-14).txt', {}), ('\u7b2c\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e03\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e09\u6b21office hour(1128-1129).txt', {}), ('\u7b2c\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e5d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e8c\u6b21 Office hour\u8a0e\u8ad6\u7d00\u9304(1120-1122).txt', {}), ('\u7b2c\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516b\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e00\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e09\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e8c\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u4e94\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u516d\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u5341\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {}), ('\u7b2c\u56db\u9031\u8ab2\u7a0b\u7b46\u8a18.txt', {})]
|
|
15:29:10,274 graphrag.index.input.text INFO Found 22 files, loading 22
|
|
15:29:10,278 graphrag.index.workflows.load INFO Workflow Run Order: ['create_base_text_units', 'create_base_extracted_entities', 'create_summarized_entities', 'create_base_entity_graph', 'create_final_entities', 'create_final_nodes', 'create_final_communities', 'create_final_relationships', 'create_final_text_units', 'create_final_community_reports', 'create_base_documents', 'create_final_documents']
|
|
15:29:10,278 graphrag.index.run.run INFO Final # of rows loaded: 22
|
|
15:29:10,392 graphrag.index.run.workflow INFO dependencies for create_base_text_units: []
|
|
15:29:10,399 datashaper.workflow.workflow INFO executing verb orderby
|
|
15:29:10,411 datashaper.workflow.workflow INFO executing verb zip
|
|
15:29:10,418 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
15:29:10,432 datashaper.workflow.workflow INFO executing verb chunk
|
|
15:29:10,733 datashaper.workflow.workflow INFO executing verb select
|
|
15:29:10,741 datashaper.workflow.workflow INFO executing verb unroll
|
|
15:29:10,753 datashaper.workflow.workflow INFO executing verb rename
|
|
15:29:10,759 datashaper.workflow.workflow INFO executing verb genid
|
|
15:29:10,771 datashaper.workflow.workflow INFO executing verb unzip
|
|
15:29:10,780 datashaper.workflow.workflow INFO executing verb copy
|
|
15:29:10,787 datashaper.workflow.workflow INFO executing verb filter
|
|
15:29:10,805 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_text_units.parquet
|
|
15:29:10,989 graphrag.index.run.workflow INFO dependencies for create_base_extracted_entities: ['create_base_text_units']
|
|
15:29:10,989 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
15:29:11,49 datashaper.workflow.workflow INFO executing verb entity_extract
|
|
15:29:11,67 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
15:29:11,77 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for gpt-4-turbo-preview: TPM=0, RPM=0
|
|
15:29:11,77 graphrag.index.llm.load_llm INFO create concurrency limiter for gpt-4-turbo-preview: 25
|
|
15:29:20,228 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:20,236 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.125. input_tokens=2292, output_tokens=350
|
|
15:29:21,507 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:21,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.359000000054948. input_tokens=2292, output_tokens=281
|
|
15:29:23,134 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:23,135 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.984000000054948. input_tokens=2294, output_tokens=284
|
|
15:29:23,263 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:23,264 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.093999999924563. input_tokens=1353, output_tokens=397
|
|
15:29:23,971 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:23,972 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.797000000020489. input_tokens=2292, output_tokens=322
|
|
15:29:25,303 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:25,308 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.202999999979511. input_tokens=2292, output_tokens=467
|
|
15:29:25,660 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:25,663 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.5. input_tokens=2292, output_tokens=409
|
|
15:29:26,11 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:26,13 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 2.890999999945052. input_tokens=1162, output_tokens=5
|
|
15:29:26,401 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:26,403 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.311999999918044. input_tokens=2292, output_tokens=577
|
|
15:29:26,438 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:26,439 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.312999999965541. input_tokens=2292, output_tokens=620
|
|
15:29:26,772 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:26,773 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.640999999945052. input_tokens=2293, output_tokens=424
|
|
15:29:26,864 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:26,866 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.73499999998603. input_tokens=2292, output_tokens=638
|
|
15:29:27,151 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:27,152 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.030999999959022. input_tokens=2292, output_tokens=423
|
|
15:29:28,982 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:28,983 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.844000000040978. input_tokens=1544, output_tokens=755
|
|
15:29:30,577 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:30,578 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.43700000003446. input_tokens=2291, output_tokens=507
|
|
15:29:30,933 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:30,939 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.82799999997951. input_tokens=2293, output_tokens=604
|
|
15:29:31,645 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:31,647 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.405999999959022. input_tokens=2293, output_tokens=472
|
|
15:29:32,162 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:32,165 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.01500000001397. input_tokens=2291, output_tokens=634
|
|
15:29:32,195 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:32,197 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.07799999997951. input_tokens=2292, output_tokens=629
|
|
15:29:32,465 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:32,468 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.29700000002049. input_tokens=2292, output_tokens=891
|
|
15:29:33,183 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:33,185 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 6.422000000020489. input_tokens=1326, output_tokens=247
|
|
15:29:33,596 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:33,605 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.469000000040978. input_tokens=2292, output_tokens=606
|
|
15:29:33,710 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:33,713 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.54700000002049. input_tokens=2293, output_tokens=817
|
|
15:29:33,927 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:33,929 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.766000000061467. input_tokens=2292, output_tokens=563
|
|
15:29:35,672 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:35,674 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.5. input_tokens=2292, output_tokens=567
|
|
15:29:36,135 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:36,137 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.968999999924563. input_tokens=2294, output_tokens=797
|
|
15:29:36,778 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:36,779 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.280999999959022. input_tokens=2260, output_tokens=491
|
|
15:29:38,832 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:38,834 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.67200000002049. input_tokens=2293, output_tokens=804
|
|
15:29:41,385 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:41,390 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.25. input_tokens=34, output_tokens=121
|
|
15:29:41,465 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:41,471 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.5. input_tokens=2293, output_tokens=453
|
|
15:29:41,655 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:41,656 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.952999999979511. input_tokens=2293, output_tokens=196
|
|
15:29:42,568 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:42,577 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.42200000002049. input_tokens=2292, output_tokens=885
|
|
15:29:43,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:43,573 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.54700000002049. input_tokens=2293, output_tokens=508
|
|
15:29:44,104 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:44,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.329000000027008. input_tokens=34, output_tokens=183
|
|
15:29:45,579 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:45,580 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.108999999938533. input_tokens=1453, output_tokens=289
|
|
15:29:46,602 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:46,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.031000000075437. input_tokens=34, output_tokens=84
|
|
15:29:48,80 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:48,87 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.5. input_tokens=2292, output_tokens=686
|
|
15:29:49,527 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:49,530 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.608999999938533. input_tokens=2294, output_tokens=438
|
|
15:29:49,919 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:49,921 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.766000000061467. input_tokens=2293, output_tokens=784
|
|
15:29:50,682 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:50,683 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.82799999997951. input_tokens=2292, output_tokens=593
|
|
15:29:50,755 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:50,757 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.56200000003446. input_tokens=2292, output_tokens=574
|
|
15:29:51,268 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:51,271 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.60999999998603. input_tokens=34, output_tokens=228
|
|
15:29:51,396 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:51,398 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.562999999965541. input_tokens=34, output_tokens=165
|
|
15:29:51,658 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:51,661 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 0.9059999999590218. input_tokens=34, output_tokens=5
|
|
15:29:51,686 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:51,688 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.25. input_tokens=2293, output_tokens=894
|
|
15:29:51,938 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:51,939 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.67200000002049. input_tokens=2292, output_tokens=713
|
|
15:29:52,50 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:52,51 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.891000000061467. input_tokens=2293, output_tokens=645
|
|
15:29:52,328 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:52,335 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.39000000001397. input_tokens=2292, output_tokens=748
|
|
15:29:52,387 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:52,388 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.85999999998603. input_tokens=34, output_tokens=77
|
|
15:29:52,638 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:52,638 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.562999999965541. input_tokens=34, output_tokens=110
|
|
15:29:52,979 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:52,981 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.375. input_tokens=2292, output_tokens=510
|
|
15:29:53,75 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:53,81 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.09299999999348. input_tokens=2292, output_tokens=849
|
|
15:29:53,550 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:53,553 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.234000000054948. input_tokens=2292, output_tokens=1001
|
|
15:29:54,288 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:54,289 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.686999999918044. input_tokens=34, output_tokens=236
|
|
15:29:55,558 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:55,561 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.98499999998603. input_tokens=34, output_tokens=236
|
|
15:29:55,773 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:55,774 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.10999999998603. input_tokens=34, output_tokens=127
|
|
15:29:55,817 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:55,818 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.422000000020489. input_tokens=34, output_tokens=360
|
|
15:29:55,872 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:55,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.234000000054948. input_tokens=34, output_tokens=106
|
|
15:29:56,124 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:56,126 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.655999999959022. input_tokens=34, output_tokens=429
|
|
15:29:56,824 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:56,829 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.71799999999348. input_tokens=34, output_tokens=367
|
|
15:29:57,572 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:57,574 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.030999999959022. input_tokens=34, output_tokens=103
|
|
15:29:58,259 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:58,261 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.875. input_tokens=34, output_tokens=91
|
|
15:29:58,595 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:58,598 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.906000000075437. input_tokens=34, output_tokens=291
|
|
15:29:58,745 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:58,747 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.702999999979511. input_tokens=34, output_tokens=237
|
|
15:29:58,826 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:58,827 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.562000000034459. input_tokens=34, output_tokens=191
|
|
15:29:59,155 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:59,156 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.57799999997951. input_tokens=34, output_tokens=415
|
|
15:29:59,428 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:29:59,429 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.5. input_tokens=34, output_tokens=223
|
|
15:30:00,430 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:00,432 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.031000000075437. input_tokens=34, output_tokens=299
|
|
15:30:00,845 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:00,847 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.078000000095926. input_tokens=34, output_tokens=110
|
|
15:30:01,228 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:01,234 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.35999999998603. input_tokens=34, output_tokens=102
|
|
15:30:02,158 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:02,159 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.592999999993481. input_tokens=34, output_tokens=127
|
|
15:30:02,293 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:02,295 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 36.64100000006147. input_tokens=2291, output_tokens=954
|
|
15:30:02,676 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:02,678 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.687000000034459. input_tokens=34, output_tokens=278
|
|
15:30:03,327 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:03,329 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.14000000001397. input_tokens=2292, output_tokens=798
|
|
15:30:03,389 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:03,391 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.718999999924563. input_tokens=1779, output_tokens=714
|
|
15:30:04,418 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:04,424 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 38.01600000006147. input_tokens=2293, output_tokens=1004
|
|
15:30:05,668 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:05,670 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 34.03100000007544. input_tokens=2292, output_tokens=887
|
|
15:30:06,858 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:06,860 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.03200000000652. input_tokens=34, output_tokens=277
|
|
15:30:07,50 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:07,52 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.109000000054948. input_tokens=34, output_tokens=186
|
|
15:30:07,109 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:07,111 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.28200000000652. input_tokens=34, output_tokens=276
|
|
15:30:07,315 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:07,317 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.077999999979511. input_tokens=34, output_tokens=126
|
|
15:30:07,466 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:07,468 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.313000000081956. input_tokens=34, output_tokens=245
|
|
15:30:08,629 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:08,635 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.469000000040978. input_tokens=34, output_tokens=155
|
|
15:30:08,828 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:08,830 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.562000000034459. input_tokens=34, output_tokens=395
|
|
15:30:10,267 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:10,268 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.155999999959022. input_tokens=1137, output_tokens=5
|
|
15:30:11,152 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:11,154 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.07799999997951. input_tokens=34, output_tokens=491
|
|
15:30:12,595 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:12,597 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 20.906000000075437. input_tokens=34, output_tokens=802
|
|
15:30:12,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:12,802 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.67200000002049. input_tokens=2292, output_tokens=430
|
|
15:30:12,983 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:12,984 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.40700000000652. input_tokens=2292, output_tokens=582
|
|
15:30:13,101 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:13,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.35999999998603. input_tokens=34, output_tokens=361
|
|
15:30:13,635 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:13,637 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.82799999997951. input_tokens=34, output_tokens=620
|
|
15:30:14,624 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:14,626 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.344000000040978. input_tokens=2292, output_tokens=536
|
|
15:30:14,763 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:14,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.343999999924563. input_tokens=2294, output_tokens=475
|
|
15:30:16,6 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:16,7 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 3.01500000001397. input_tokens=1272, output_tokens=5
|
|
15:30:16,712 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:16,713 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.280999999959022. input_tokens=2292, output_tokens=662
|
|
15:30:17,959 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:17,960 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.530999999959022. input_tokens=34, output_tokens=400
|
|
15:30:18,975 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:18,977 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 20.375. input_tokens=34, output_tokens=564
|
|
15:30:19,592 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:19,593 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 27.266000000061467. input_tokens=34, output_tokens=912
|
|
15:30:20,140 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:20,142 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.81299999996554. input_tokens=2291, output_tokens=606
|
|
15:30:20,527 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:20,535 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.21799999999348. input_tokens=2292, output_tokens=503
|
|
15:30:20,745 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:20,747 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.280999999959022. input_tokens=2293, output_tokens=467
|
|
15:30:22,288 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:22,290 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.452999999979511. input_tokens=2291, output_tokens=393
|
|
15:30:22,814 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:22,818 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.952999999979511. input_tokens=34, output_tokens=439
|
|
15:30:22,833 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:22,835 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.687000000034459. input_tokens=34, output_tokens=76
|
|
15:30:23,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:23,859 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.18799999996554. input_tokens=34, output_tokens=510
|
|
15:30:23,911 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:23,913 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.64000000001397. input_tokens=2292, output_tokens=480
|
|
15:30:24,735 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:24,737 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 22.06299999996554. input_tokens=34, output_tokens=636
|
|
15:30:25,476 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:25,477 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.5. input_tokens=34, output_tokens=209
|
|
15:30:26,108 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:26,111 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 23.81299999996554. input_tokens=34, output_tokens=761
|
|
15:30:26,316 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:26,317 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.562999999965541. input_tokens=34, output_tokens=159
|
|
15:30:27,105 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:27,111 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.0. input_tokens=2292, output_tokens=455
|
|
15:30:27,390 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:27,392 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.0. input_tokens=2292, output_tokens=721
|
|
15:30:28,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:28,686 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.15700000000652. input_tokens=34, output_tokens=234
|
|
15:30:28,971 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:29,191 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.343999999924563. input_tokens=2293, output_tokens=947
|
|
15:30:30,743 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:30,745 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.594000000040978. input_tokens=2293, output_tokens=498
|
|
15:30:30,771 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:30,894 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.890999999945052. input_tokens=2292, output_tokens=444
|
|
15:30:31,151 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:31,151 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.436999999918044. input_tokens=2294, output_tokens=457
|
|
15:30:31,267 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:31,274 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.640999999945052. input_tokens=2292, output_tokens=604
|
|
15:30:31,844 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:31,846 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.203000000095926. input_tokens=2292, output_tokens=563
|
|
15:30:32,56 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:32,58 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.281000000075437. input_tokens=2293, output_tokens=545
|
|
15:30:32,172 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:32,174 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.57799999997951. input_tokens=2292, output_tokens=617
|
|
15:30:32,646 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:33,110 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:33,112 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.797000000020489. input_tokens=34, output_tokens=148
|
|
15:30:33,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.45299999997951. input_tokens=2237, output_tokens=666
|
|
15:30:36,89 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:36,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.281000000075437. input_tokens=2293, output_tokens=470
|
|
15:30:36,527 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:36,529 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.905999999959022. input_tokens=2293, output_tokens=569
|
|
15:30:37,750 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:37,752 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.562000000034459. input_tokens=34, output_tokens=203
|
|
15:30:38,825 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:38,829 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.0. input_tokens=2293, output_tokens=634
|
|
15:30:39,15 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:39,17 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.155999999959022. input_tokens=2292, output_tokens=610
|
|
15:30:39,157 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:39,158 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.092999999993481. input_tokens=34, output_tokens=164
|
|
15:30:39,786 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:39,793 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.0. input_tokens=2292, output_tokens=725
|
|
15:30:39,845 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:39,848 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.453000000095926. input_tokens=1388, output_tokens=381
|
|
15:30:40,0 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:40,2 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.312000000034459. input_tokens=34, output_tokens=382
|
|
15:30:40,586 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:40,587 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.733999999938533. input_tokens=34, output_tokens=166
|
|
15:30:42,990 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:42,992 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.890999999945052. input_tokens=34, output_tokens=245
|
|
15:30:43,587 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:43,588 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.827999999979511. input_tokens=34, output_tokens=199
|
|
15:30:43,671 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:43,673 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.93700000003446. input_tokens=2291, output_tokens=497
|
|
15:30:44,356 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:44,357 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.25. input_tokens=2292, output_tokens=584
|
|
15:30:45,265 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:45,267 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.10999999998603. input_tokens=34, output_tokens=379
|
|
15:30:45,577 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:45,580 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.625. input_tokens=2292, output_tokens=1042
|
|
15:30:46,144 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:46,152 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.625. input_tokens=34, output_tokens=276
|
|
15:30:46,601 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:46,605 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.85999999998603. input_tokens=2292, output_tokens=508
|
|
15:30:46,755 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:46,756 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.280999999959022. input_tokens=2292, output_tokens=727
|
|
15:30:47,225 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:47,228 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 24.938000000081956. input_tokens=34, output_tokens=699
|
|
15:30:48,112 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:48,114 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.28200000000652. input_tokens=34, output_tokens=330
|
|
15:30:48,148 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:48,151 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.030999999959022. input_tokens=2292, output_tokens=415
|
|
15:30:49,471 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:49,472 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 29.875. input_tokens=2292, output_tokens=815
|
|
15:30:49,609 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:49,610 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.937999999965541. input_tokens=34, output_tokens=126
|
|
15:30:50,25 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:50,27 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.125. input_tokens=2292, output_tokens=500
|
|
15:30:50,950 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:50,952 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.04700000002049. input_tokens=2292, output_tokens=715
|
|
15:30:51,22 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:51,22 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.405999999959022. input_tokens=34, output_tokens=86
|
|
15:30:52,338 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:52,340 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.188000000081956. input_tokens=34, output_tokens=170
|
|
15:30:53,52 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:53,54 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.93700000003446. input_tokens=2292, output_tokens=706
|
|
15:30:53,220 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:53,222 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.078000000095926. input_tokens=34, output_tokens=119
|
|
15:30:54,558 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:54,559 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.093999999924563. input_tokens=34, output_tokens=216
|
|
15:30:55,679 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:55,775 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.187999999965541. input_tokens=2019, output_tokens=434
|
|
15:30:56,68 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:56,69 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.702999999979511. input_tokens=34, output_tokens=396
|
|
15:30:57,370 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:57,373 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.219000000040978. input_tokens=2293, output_tokens=466
|
|
15:30:57,895 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:57,896 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.093999999924563. input_tokens=2293, output_tokens=649
|
|
15:30:57,900 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:57,903 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.920999999972992. input_tokens=2292, output_tokens=527
|
|
15:30:58,140 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:58,143 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.640999999945052. input_tokens=2293, output_tokens=657
|
|
15:30:58,205 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:58,205 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.983999999938533. input_tokens=34, output_tokens=123
|
|
15:30:58,736 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:58,737 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.687999999965541. input_tokens=34, output_tokens=179
|
|
15:30:59,642 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:30:59,645 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.890999999945052. input_tokens=34, output_tokens=474
|
|
15:31:00,10 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:00,11 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.98499999998603. input_tokens=34, output_tokens=229
|
|
15:31:00,51 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:00,52 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.094000000040978. input_tokens=34, output_tokens=221
|
|
15:31:01,70 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:01,71 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.06299999996554. input_tokens=2293, output_tokens=584
|
|
15:31:01,360 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:01,361 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.297000000020489. input_tokens=34, output_tokens=145
|
|
15:31:01,599 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:01,600 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.484000000054948. input_tokens=34, output_tokens=358
|
|
15:31:02,466 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:02,467 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.703000000095926. input_tokens=34, output_tokens=286
|
|
15:31:03,157 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:03,159 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 30.983999999938533. input_tokens=2292, output_tokens=889
|
|
15:31:03,923 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:03,925 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.906000000075437. input_tokens=34, output_tokens=130
|
|
15:31:06,711 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:06,713 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.125. input_tokens=2293, output_tokens=659
|
|
15:31:06,863 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:06,864 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.5. input_tokens=34, output_tokens=182
|
|
15:31:07,305 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:07,306 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.234000000054948. input_tokens=34, output_tokens=176
|
|
15:31:07,785 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:07,788 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.8589999999385327. input_tokens=34, output_tokens=121
|
|
15:31:08,339 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:08,341 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 22.766000000061467. input_tokens=34, output_tokens=550
|
|
15:31:08,569 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:08,580 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.375. input_tokens=34, output_tokens=253
|
|
15:31:09,14 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:09,16 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.875. input_tokens=1315, output_tokens=229
|
|
15:31:09,263 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:09,264 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.25. input_tokens=34, output_tokens=488
|
|
15:31:10,258 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:10,259 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.234000000054948. input_tokens=2292, output_tokens=915
|
|
15:31:11,249 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:11,252 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 31.405999999959022. input_tokens=34, output_tokens=845
|
|
15:31:11,502 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:11,504 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.452999999979511. input_tokens=2292, output_tokens=459
|
|
15:31:12,243 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:12,246 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.344000000040978. input_tokens=2291, output_tokens=531
|
|
15:31:13,463 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:13,466 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.609000000054948. input_tokens=34, output_tokens=187
|
|
15:31:14,635 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:14,637 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.905999999959022. input_tokens=34, output_tokens=415
|
|
15:31:15,130 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:15,132 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.827999999979511. input_tokens=34, output_tokens=252
|
|
15:31:16,74 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:16,87 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.51500000001397. input_tokens=2291, output_tokens=576
|
|
15:31:16,949 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:16,950 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.452999999979511. input_tokens=34, output_tokens=132
|
|
15:31:17,91 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:17,93 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.625. input_tokens=34, output_tokens=107
|
|
15:31:17,249 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:17,266 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:17,269 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.921999999904074. input_tokens=2293, output_tokens=1001
|
|
15:31:17,269 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.375. input_tokens=2290, output_tokens=501
|
|
15:31:18,843 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:18,844 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.25. input_tokens=2294, output_tokens=507
|
|
15:31:19,212 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:19,213 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.952999999979511. input_tokens=34, output_tokens=221
|
|
15:31:19,438 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:19,440 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.85999999998603. input_tokens=34, output_tokens=267
|
|
15:31:21,69 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:21,70 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.90700000000652. input_tokens=34, output_tokens=687
|
|
15:31:21,515 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:21,521 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.875. input_tokens=2292, output_tokens=627
|
|
15:31:22,508 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:22,511 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 35.28099999995902. input_tokens=34, output_tokens=954
|
|
15:31:22,745 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:22,747 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.625. input_tokens=2294, output_tokens=252
|
|
15:31:24,174 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:24,176 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.70299999997951. input_tokens=2292, output_tokens=512
|
|
15:31:25,77 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:25,80 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 53.81200000003446. input_tokens=34, output_tokens=1506
|
|
15:31:25,188 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:25,190 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.015999999945052. input_tokens=1122, output_tokens=5
|
|
15:31:26,21 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:26,21 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.75. input_tokens=34, output_tokens=294
|
|
15:31:27,431 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:27,432 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 37.81200000003446. input_tokens=2293, output_tokens=1071
|
|
15:31:30,385 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:30,387 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.952999999979511. input_tokens=34, output_tokens=5
|
|
15:31:30,504 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:30,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.234000000054948. input_tokens=2293, output_tokens=834
|
|
15:31:31,31 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:31,32 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.686999999918044. input_tokens=2293, output_tokens=427
|
|
15:31:31,718 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:31,721 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 46.453000000095926. input_tokens=2292, output_tokens=1216
|
|
15:31:31,944 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:31,948 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.5. input_tokens=2293, output_tokens=437
|
|
15:31:32,588 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:32,591 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.531000000075437. input_tokens=34, output_tokens=383
|
|
15:31:32,699 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:32,703 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.0. input_tokens=2182, output_tokens=517
|
|
15:31:33,41 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:33,45 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.79700000002049. input_tokens=2292, output_tokens=586
|
|
15:31:35,363 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:35,365 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.515999999945052. input_tokens=2292, output_tokens=403
|
|
15:31:36,758 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:36,763 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.046999999904074. input_tokens=34, output_tokens=136
|
|
15:31:37,106 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:37,109 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.015999999945052. input_tokens=2292, output_tokens=847
|
|
15:31:37,218 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:37,220 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.141000000061467. input_tokens=2293, output_tokens=506
|
|
15:31:37,952 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:37,956 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.437000000034459. input_tokens=34, output_tokens=625
|
|
15:31:38,931 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:38,933 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.906000000075437. input_tokens=2292, output_tokens=490
|
|
15:31:39,244 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:39,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.469000000040978. input_tokens=2292, output_tokens=573
|
|
15:31:41,276 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:41,278 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.01500000001397. input_tokens=2293, output_tokens=625
|
|
15:31:41,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:41,293 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.54700000002049. input_tokens=2221, output_tokens=692
|
|
15:31:41,813 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:41,819 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.312999999965541. input_tokens=34, output_tokens=403
|
|
15:31:42,540 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:42,560 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 33.54700000002049. input_tokens=2292, output_tokens=582
|
|
15:31:42,648 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:42,649 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.046999999904074. input_tokens=2292, output_tokens=356
|
|
15:31:43,14 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:43,16 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 45.64099999994505. input_tokens=2292, output_tokens=1078
|
|
15:31:43,651 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:43,653 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.70299999997951. input_tokens=2292, output_tokens=663
|
|
15:31:44,556 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:44,559 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.202999999979511. input_tokens=34, output_tokens=273
|
|
15:31:45,122 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:45,124 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.92200000002049. input_tokens=2292, output_tokens=700
|
|
15:31:45,720 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:45,729 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.672000000020489. input_tokens=2293, output_tokens=480
|
|
15:31:45,806 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:45,809 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.29700000002049. input_tokens=2293, output_tokens=655
|
|
15:31:46,760 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:46,769 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.812999999965541. input_tokens=34, output_tokens=384
|
|
15:31:47,23 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:47,25 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.2969999999040738. input_tokens=34, output_tokens=5
|
|
15:31:48,168 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:48,174 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.094000000040978. input_tokens=2292, output_tokens=740
|
|
15:31:48,209 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:48,210 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.39000000001397. input_tokens=34, output_tokens=148
|
|
15:31:48,706 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:48,707 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.687000000034459. input_tokens=34, output_tokens=172
|
|
15:31:48,900 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:48,904 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.51500000001397. input_tokens=2292, output_tokens=582
|
|
15:31:49,318 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:49,319 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 37.06299999996554. input_tokens=34, output_tokens=600
|
|
15:31:50,344 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:50,346 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 35.703000000095926. input_tokens=34, output_tokens=952
|
|
15:31:51,202 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:51,203 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.25. input_tokens=34, output_tokens=567
|
|
15:31:51,906 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:51,908 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.342999999993481. input_tokens=34, output_tokens=139
|
|
15:31:52,169 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:52,187 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.422000000020489. input_tokens=2293, output_tokens=554
|
|
15:31:52,837 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:52,840 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.813000000081956. input_tokens=2292, output_tokens=622
|
|
15:31:52,947 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:52,955 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.312000000034459. input_tokens=34, output_tokens=261
|
|
15:31:53,258 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:53,263 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.9219999999040738. input_tokens=34, output_tokens=72
|
|
15:31:54,20 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:54,22 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.82799999997951. input_tokens=2291, output_tokens=716
|
|
15:31:55,356 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:55,369 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.65700000000652. input_tokens=2292, output_tokens=1021
|
|
15:31:58,555 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:58,557 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.43700000003446. input_tokens=2292, output_tokens=646
|
|
15:31:58,637 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:58,639 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.280999999959022. input_tokens=34, output_tokens=82
|
|
15:31:59,481 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:31:59,483 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 20.54700000002049. input_tokens=34, output_tokens=560
|
|
15:32:00,884 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:00,886 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.23499999998603. input_tokens=2292, output_tokens=368
|
|
15:32:00,905 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:00,912 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.89000000001397. input_tokens=34, output_tokens=356
|
|
15:32:01,170 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:01,173 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.906000000075437. input_tokens=34, output_tokens=210
|
|
15:32:02,402 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:02,403 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.125. input_tokens=1807, output_tokens=474
|
|
15:32:02,928 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:02,930 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.359000000054948. input_tokens=2293, output_tokens=783
|
|
15:32:03,37 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:03,41 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.780999999959022. input_tokens=2292, output_tokens=634
|
|
15:32:03,472 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:03,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.344000000040978. input_tokens=1832, output_tokens=702
|
|
15:32:04,353 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:04,355 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.4380000000819564. input_tokens=34, output_tokens=82
|
|
15:32:04,745 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:04,747 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.562000000034459. input_tokens=34, output_tokens=254
|
|
15:32:04,764 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:04,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.06299999996554. input_tokens=2293, output_tokens=437
|
|
15:32:04,842 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:04,847 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.67200000002049. input_tokens=34, output_tokens=426
|
|
15:32:05,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:05,358 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.344000000040978. input_tokens=34, output_tokens=280
|
|
15:32:05,363 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:05,364 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.437999999965541. input_tokens=34, output_tokens=78
|
|
15:32:06,835 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:06,836 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.922000000020489. input_tokens=34, output_tokens=473
|
|
15:32:07,350 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:07,352 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 26.04700000002049. input_tokens=2292, output_tokens=710
|
|
15:32:07,356 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:07,358 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.54700000002049. input_tokens=2292, output_tokens=573
|
|
15:32:08,954 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:08,961 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.46799999999348. input_tokens=34, output_tokens=250
|
|
15:32:09,379 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:09,380 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.17200000002049. input_tokens=2292, output_tokens=481
|
|
15:32:09,872 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:09,874 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.827999999979511. input_tokens=34, output_tokens=222
|
|
15:32:10,21 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:10,23 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.06299999996554. input_tokens=34, output_tokens=237
|
|
15:32:10,88 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:10,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.453000000095926. input_tokens=34, output_tokens=261
|
|
15:32:10,659 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:10,660 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.25. input_tokens=34, output_tokens=331
|
|
15:32:11,732 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:11,734 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.890999999945052. input_tokens=2293, output_tokens=798
|
|
15:32:11,888 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:11,889 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 7.125. input_tokens=2292, output_tokens=307
|
|
15:32:12,164 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:12,166 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 23.95299999997951. input_tokens=34, output_tokens=660
|
|
15:32:12,449 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:12,451 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.092999999993481. input_tokens=34, output_tokens=217
|
|
15:32:12,649 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:12,651 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 23.32799999997951. input_tokens=34, output_tokens=628
|
|
15:32:12,690 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:12,691 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.593999999924563. input_tokens=34, output_tokens=75
|
|
15:32:12,732 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:12,744 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 23.82900000002701. input_tokens=34, output_tokens=724
|
|
15:32:13,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:13,571 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.9070000000065193. input_tokens=34, output_tokens=71
|
|
15:32:14,411 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:14,412 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 9.561999999918044. input_tokens=2292, output_tokens=274
|
|
15:32:15,49 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:15,50 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.672000000020489. input_tokens=34, output_tokens=150
|
|
15:32:15,408 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:15,411 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.452999999979511. input_tokens=34, output_tokens=162
|
|
15:32:15,509 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:15,512 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.155999999959022. input_tokens=2292, output_tokens=440
|
|
15:32:17,315 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:17,317 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.42200000002049. input_tokens=34, output_tokens=677
|
|
15:32:18,642 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:18,645 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 31.875. input_tokens=34, output_tokens=750
|
|
15:32:18,742 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:18,743 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.56299999996554. input_tokens=2293, output_tokens=473
|
|
15:32:20,485 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:20,486 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.65700000000652. input_tokens=2292, output_tokens=397
|
|
15:32:21,153 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:21,155 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.420999999972992. input_tokens=34, output_tokens=318
|
|
15:32:21,187 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:21,190 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 16.43799999996554. input_tokens=34, output_tokens=238
|
|
15:32:21,204 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:21,207 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.75. input_tokens=34, output_tokens=205
|
|
15:32:22,621 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:22,622 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.89000000001397. input_tokens=34, output_tokens=274
|
|
15:32:22,654 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:22,654 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.780999999959022. input_tokens=2292, output_tokens=406
|
|
15:32:22,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:22,723 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.406000000075437. input_tokens=34, output_tokens=117
|
|
15:32:23,709 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:23,710 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.14000000001397. input_tokens=34, output_tokens=338
|
|
15:32:24,205 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:24,207 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.733999999938533. input_tokens=2292, output_tokens=572
|
|
15:32:25,105 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:25,121 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.76500000001397. input_tokens=2066, output_tokens=631
|
|
15:32:26,350 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:26,351 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.156000000075437. input_tokens=34, output_tokens=167
|
|
15:32:27,452 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:27,454 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.25. input_tokens=34, output_tokens=264
|
|
15:32:27,562 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:27,568 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.20299999997951. input_tokens=2292, output_tokens=565
|
|
15:32:28,757 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:28,758 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 30.20299999997951. input_tokens=2292, output_tokens=1185
|
|
15:32:29,378 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:29,379 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.719000000040978. input_tokens=34, output_tokens=277
|
|
15:32:29,397 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:29,399 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.98499999998603. input_tokens=34, output_tokens=530
|
|
15:32:31,115 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:31,117 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 12.469000000040978. input_tokens=2292, output_tokens=572
|
|
15:32:31,887 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:31,890 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.546999999904074. input_tokens=34, output_tokens=166
|
|
15:32:31,994 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:31,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.532000000006519. input_tokens=34, output_tokens=110
|
|
15:32:32,502 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:32,504 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.344000000040978. input_tokens=2292, output_tokens=576
|
|
15:32:33,236 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:33,237 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.5. input_tokens=2292, output_tokens=662
|
|
15:32:33,441 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:33,444 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 18.03200000000652. input_tokens=2292, output_tokens=650
|
|
15:32:33,551 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:33,554 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.859000000054948. input_tokens=2292, output_tokens=801
|
|
15:32:35,871 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:35,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.984000000054948. input_tokens=2293, output_tokens=677
|
|
15:32:36,486 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:36,487 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 8.922000000020489. input_tokens=1819, output_tokens=330
|
|
15:32:36,753 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:36,754 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 29.405999999959022. input_tokens=2291, output_tokens=838
|
|
15:32:37,801 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:37,802 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.359000000054948. input_tokens=34, output_tokens=92
|
|
15:32:37,827 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:37,829 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 1.952999999979511. input_tokens=1110, output_tokens=44
|
|
15:32:40,122 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:40,124 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.75. input_tokens=2292, output_tokens=361
|
|
15:32:40,330 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:40,333 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.34299999999348. input_tokens=34, output_tokens=275
|
|
15:32:43,875 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:43,883 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.17200000002049. input_tokens=2294, output_tokens=639
|
|
15:32:43,966 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:43,968 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.422000000020489. input_tokens=2212, output_tokens=457
|
|
15:32:44,139 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:44,329 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.67200000002049. input_tokens=2291, output_tokens=676
|
|
15:32:44,630 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:44,633 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.0. input_tokens=2292, output_tokens=500
|
|
15:32:44,891 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:44,895 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 29.843999999924563. input_tokens=34, output_tokens=1058
|
|
15:32:47,341 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:47,360 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.15700000000652. input_tokens=2293, output_tokens=825
|
|
15:32:48,33 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:48,34 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.545999999972992. input_tokens=34, output_tokens=362
|
|
15:32:48,389 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:48,392 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 23.265999999945052. input_tokens=2291, output_tokens=586
|
|
15:32:49,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:49,321 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.687999999965541. input_tokens=34, output_tokens=141
|
|
15:32:49,497 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:49,503 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.109000000054948. input_tokens=2294, output_tokens=679
|
|
15:32:49,841 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:49,843 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 17.344000000040978. input_tokens=2292, output_tokens=539
|
|
15:32:49,902 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:50,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:50,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 5.969000000040978. input_tokens=1184, output_tokens=70
|
|
15:32:50,792 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:50,794 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 28.07799999997951. input_tokens=2292, output_tokens=826
|
|
15:32:51,38 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:51,41 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 22.280999999959022. input_tokens=2292, output_tokens=678
|
|
15:32:51,330 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:51,350 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 30.188000000081956. input_tokens=2293, output_tokens=778
|
|
15:32:51,437 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:51,441 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 41.42200000002049. input_tokens=2291, output_tokens=1180
|
|
15:32:52,89 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:52,90 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.344000000040978. input_tokens=34, output_tokens=299
|
|
15:32:52,110 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:52,111 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 14.312999999965541. input_tokens=2292, output_tokens=445
|
|
15:32:54,335 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:54,341 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 38.828000000095926. input_tokens=1748, output_tokens=1082
|
|
15:32:54,624 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.75. input_tokens=34, output_tokens=199
|
|
15:32:54,897 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:54,899 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.655999999959022. input_tokens=2292, output_tokens=439
|
|
15:32:55,38 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:55,41 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.733999999938533. input_tokens=34, output_tokens=79
|
|
15:32:55,187 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:55,189 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 24.07799999997951. input_tokens=2293, output_tokens=829
|
|
15:32:55,772 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:55,774 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.875. input_tokens=34, output_tokens=289
|
|
15:32:57,36 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:57,39 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 25.14000000001397. input_tokens=2292, output_tokens=792
|
|
15:32:58,646 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:58,648 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.015999999945052. input_tokens=34, output_tokens=121
|
|
15:32:59,286 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:59,289 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 38.79599999997299. input_tokens=2291, output_tokens=951
|
|
15:32:59,600 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:32:59,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 82.375. input_tokens=34, output_tokens=2968
|
|
15:33:00,257 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:00,261 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.875. input_tokens=34, output_tokens=325
|
|
15:33:00,282 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:00,283 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.84299999999348. input_tokens=34, output_tokens=264
|
|
15:33:00,391 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:00,393 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.265999999945052. input_tokens=2292, output_tokens=799
|
|
15:33:00,934 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:00,935 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.843999999924563. input_tokens=34, output_tokens=214
|
|
15:33:01,612 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:01,613 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.28200000000652. input_tokens=2292, output_tokens=550
|
|
15:33:02,114 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:02,115 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.765999999945052. input_tokens=34, output_tokens=379
|
|
15:33:02,930 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:02,932 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 15.562000000034459. input_tokens=2291, output_tokens=502
|
|
15:33:03,637 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:03,638 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.530999999959022. input_tokens=34, output_tokens=443
|
|
15:33:03,686 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:03,689 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.297000000020489. input_tokens=34, output_tokens=69
|
|
15:33:03,723 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:03,727 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.953000000095926. input_tokens=34, output_tokens=220
|
|
15:33:03,858 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:03,867 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.515999999945052. input_tokens=34, output_tokens=302
|
|
15:33:04,903 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:04,905 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.71799999999348. input_tokens=34, output_tokens=219
|
|
15:33:05,435 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:05,439 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.468999999924563. input_tokens=2284, output_tokens=782
|
|
15:33:06,384 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:06,384 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.687000000034459. input_tokens=34, output_tokens=5
|
|
15:33:06,531 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:06,536 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 2.8119999999180436. input_tokens=34, output_tokens=72
|
|
15:33:06,840 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:06,841 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 11.813000000081956. input_tokens=2291, output_tokens=418
|
|
15:33:08,487 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:08,488 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 6.875. input_tokens=34, output_tokens=213
|
|
15:33:08,518 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:08,519 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 1.6719999999040738. input_tokens=34, output_tokens=5
|
|
15:33:09,263 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:09,266 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 0.7809999999590218. input_tokens=1098, output_tokens=5
|
|
15:33:09,363 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:09,364 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 19.515999999945052. input_tokens=2293, output_tokens=742
|
|
15:33:09,688 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:09,690 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 31.85999999998603. input_tokens=2291, output_tokens=887
|
|
15:33:10,449 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:10,452 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 15.562000000034459. input_tokens=34, output_tokens=602
|
|
15:33:10,889 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:10,890 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 10.625. input_tokens=34, output_tokens=367
|
|
15:33:11,486 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:11,487 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 21.98499999998603. input_tokens=2292, output_tokens=741
|
|
15:33:12,155 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:12,157 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.875. input_tokens=34, output_tokens=431
|
|
15:33:12,974 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:12,976 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 10.859000000054948. input_tokens=2292, output_tokens=342
|
|
15:33:13,80 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:13,81 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 14.437000000034459. input_tokens=34, output_tokens=485
|
|
15:33:14,57 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:14,58 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 4.702999999979511. input_tokens=34, output_tokens=126
|
|
15:33:14,368 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:14,370 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.422000000020489. input_tokens=34, output_tokens=363
|
|
15:33:14,476 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:14,481 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.954000000027008. input_tokens=34, output_tokens=311
|
|
15:33:14,687 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:14,688 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.797000000020489. input_tokens=34, output_tokens=117
|
|
15:33:15,778 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:15,780 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 27.75. input_tokens=2292, output_tokens=940
|
|
15:33:15,826 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:15,829 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.187000000034459. input_tokens=34, output_tokens=408
|
|
15:33:16,182 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:16,185 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 11.28200000000652. input_tokens=34, output_tokens=217
|
|
15:33:16,217 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:16,218 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.766000000061467. input_tokens=34, output_tokens=172
|
|
15:33:16,621 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:16,624 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.76500000001397. input_tokens=34, output_tokens=408
|
|
15:33:16,743 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:16,745 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 3.6570000000065193. input_tokens=34, output_tokens=93
|
|
15:33:17,98 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:17,99 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 16.813000000081956. input_tokens=2292, output_tokens=582
|
|
15:33:17,305 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:17,306 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 17.70299999997951. input_tokens=34, output_tokens=697
|
|
15:33:17,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:17,784 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.75. input_tokens=2292, output_tokens=420
|
|
15:33:19,428 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:19,428 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.062000000034459. input_tokens=34, output_tokens=110
|
|
15:33:19,827 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:19,830 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 13.452999999979511. input_tokens=2198, output_tokens=487
|
|
15:33:20,205 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:20,207 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 29.17200000002049. input_tokens=1923, output_tokens=825
|
|
15:33:20,325 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:20,326 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.172000000020489. input_tokens=34, output_tokens=270
|
|
15:33:20,994 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:21,7 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.51500000001397. input_tokens=34, output_tokens=238
|
|
15:33:21,233 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:21,236 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 20.29700000002049. input_tokens=2293, output_tokens=532
|
|
15:33:22,286 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:22,288 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 13.76500000001397. input_tokens=34, output_tokens=497
|
|
15:33:22,360 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:22,362 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 7.875. input_tokens=34, output_tokens=313
|
|
15:33:23,14 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:23,19 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.327999999979511. input_tokens=34, output_tokens=287
|
|
15:33:23,237 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:23,238 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 9.172000000020489. input_tokens=34, output_tokens=374
|
|
15:33:23,932 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:23,936 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "Process" with 0 retries took 33.14099999994505. input_tokens=2291, output_tokens=996
|
|
15:33:25,213 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:25,216 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 12.25. input_tokens=34, output_tokens=460
|
|
15:33:25,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:25,475 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.375. input_tokens=34, output_tokens=205
|
|
15:33:25,735 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:25,737 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 5.532000000006519. input_tokens=34, output_tokens=114
|
|
15:33:29,271 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:29,275 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 8.030999999959022. input_tokens=34, output_tokens=239
|
|
15:33:35,688 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:35,689 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 26.42200000002049. input_tokens=34, output_tokens=825
|
|
15:33:36,79 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:36,80 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 18.29700000002049. input_tokens=34, output_tokens=387
|
|
15:33:41,361 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:41,368 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 52.04700000002049. input_tokens=34, output_tokens=1133
|
|
15:33:42,412 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:42,414 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 26.625. input_tokens=34, output_tokens=813
|
|
15:33:43,865 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:43,866 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 24.03200000000652. input_tokens=34, output_tokens=756
|
|
15:33:46,445 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:46,448 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 41.0. input_tokens=34, output_tokens=1238
|
|
15:33:50,106 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:50,110 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 40.42200000002049. input_tokens=34, output_tokens=1153
|
|
15:33:52,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:52,359 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "extract-continuation-0" with 0 retries took 28.42200000002049. input_tokens=34, output_tokens=890
|
|
15:33:52,378 datashaper.workflow.workflow INFO executing verb snapshot
|
|
15:33:52,417 datashaper.workflow.workflow INFO executing verb merge_graphs
|
|
15:33:52,617 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
15:33:52,624 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_extracted_entities.parquet
|
|
15:33:52,776 graphrag.index.run.workflow INFO dependencies for create_summarized_entities: ['create_base_extracted_entities']
|
|
15:33:52,777 graphrag.utils.storage INFO read table from storage: create_base_extracted_entities.parquet
|
|
15:33:52,800 datashaper.workflow.workflow INFO executing verb summarize_descriptions
|
|
15:33:56,364 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:56,366 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.375. input_tokens=157, output_tokens=60
|
|
15:33:56,406 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:56,408 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.4209999999729916. input_tokens=153, output_tokens=47
|
|
15:33:56,760 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:56,762 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.843999999924563. input_tokens=272, output_tokens=141
|
|
15:33:56,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:56,782 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.8119999999180436. input_tokens=180, output_tokens=86
|
|
15:33:57,182 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:57,187 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.202999999979511. input_tokens=177, output_tokens=85
|
|
15:33:57,785 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:57,786 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.795999999972992. input_tokens=156, output_tokens=58
|
|
15:33:59,347 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:33:59,349 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.359000000054948. input_tokens=250, output_tokens=141
|
|
15:34:00,792 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:00,794 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.031000000075437. input_tokens=147, output_tokens=70
|
|
15:34:00,836 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:00,837 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.858999999938533. input_tokens=225, output_tokens=162
|
|
15:34:02,104 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:02,105 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.15700000000652. input_tokens=226, output_tokens=167
|
|
15:34:02,729 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:02,731 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.375. input_tokens=172, output_tokens=109
|
|
15:34:02,921 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:02,922 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.516000000061467. input_tokens=189, output_tokens=147
|
|
15:34:02,938 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:02,940 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.593999999924563. input_tokens=173, output_tokens=84
|
|
15:34:03,156 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:03,159 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.186999999918044. input_tokens=1346, output_tokens=431
|
|
15:34:03,333 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:03,334 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.14000000001397. input_tokens=167, output_tokens=50
|
|
15:34:03,520 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:03,521 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.515999999945052. input_tokens=225, output_tokens=183
|
|
15:34:03,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:03,862 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.922000000020489. input_tokens=298, output_tokens=215
|
|
15:34:04,10 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:04,12 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.030999999959022. input_tokens=250, output_tokens=278
|
|
15:34:04,362 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:04,363 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.40700000000652. input_tokens=295, output_tokens=317
|
|
15:34:05,66 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:05,68 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 12.125. input_tokens=371, output_tokens=349
|
|
15:34:05,339 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:05,340 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.0160000000614673. input_tokens=175, output_tokens=57
|
|
15:34:05,694 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:05,696 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.85999999998603. input_tokens=163, output_tokens=96
|
|
15:34:06,151 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:06,160 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.233999999938533. input_tokens=381, output_tokens=379
|
|
15:34:06,461 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:06,463 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.672000000020489. input_tokens=164, output_tokens=118
|
|
15:34:07,1 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:07,3 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.047000000020489. input_tokens=239, output_tokens=258
|
|
15:34:07,325 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:07,326 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.530999999959022. input_tokens=175, output_tokens=142
|
|
15:34:07,433 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:07,434 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.5. input_tokens=455, output_tokens=440
|
|
15:34:08,979 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:08,981 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.125. input_tokens=164, output_tokens=82
|
|
15:34:09,701 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:09,704 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.187000000034459. input_tokens=245, output_tokens=171
|
|
15:34:10,7 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:10,9 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.562000000034459. input_tokens=143, output_tokens=32
|
|
15:34:11,635 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:11,637 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 18.68799999996554. input_tokens=677, output_tokens=322
|
|
15:34:11,842 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:11,843 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.5. input_tokens=244, output_tokens=127
|
|
15:34:11,967 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:11,968 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.813000000081956. input_tokens=172, output_tokens=85
|
|
15:34:12,74 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:12,80 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.155999999959022. input_tokens=362, output_tokens=364
|
|
15:34:12,271 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:12,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.812999999965541. input_tokens=250, output_tokens=156
|
|
15:34:12,565 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:12,565 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.60999999998603. input_tokens=299, output_tokens=337
|
|
15:34:12,606 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:12,608 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.640999999945052. input_tokens=446, output_tokens=392
|
|
15:34:13,834 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:13,836 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.842999999993481. input_tokens=193, output_tokens=151
|
|
15:34:13,998 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:13,999 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.844000000040978. input_tokens=320, output_tokens=189
|
|
15:34:14,341 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:14,342 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.375. input_tokens=154, output_tokens=61
|
|
15:34:14,412 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:14,414 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.39000000001397. input_tokens=658, output_tokens=500
|
|
15:34:14,930 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:14,935 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.297000000020489. input_tokens=177, output_tokens=94
|
|
15:34:15,530 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:15,532 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.530999999959022. input_tokens=245, output_tokens=106
|
|
15:34:15,632 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:15,633 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.922000000020489. input_tokens=198, output_tokens=107
|
|
15:34:16,123 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:16,124 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.125. input_tokens=219, output_tokens=135
|
|
15:34:16,639 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:16,640 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.702999999979511. input_tokens=290, output_tokens=426
|
|
15:34:17,868 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:17,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 24.95299999997951. input_tokens=452, output_tokens=500
|
|
15:34:17,929 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:17,934 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.827999999979511. input_tokens=422, output_tokens=444
|
|
15:34:18,63 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,65 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.452999999979511. input_tokens=194, output_tokens=107
|
|
15:34:18,83 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,85 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 13.01500000001397. input_tokens=193, output_tokens=143
|
|
15:34:18,247 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,249 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.327999999979511. input_tokens=309, output_tokens=452
|
|
15:34:18,408 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,409 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 25.45299999997951. input_tokens=628, output_tokens=500
|
|
15:34:18,654 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,655 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.811999999918044. input_tokens=392, output_tokens=175
|
|
15:34:18,691 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:18,693 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.562999999965541. input_tokens=144, output_tokens=51
|
|
15:34:19,248 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:19,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.922000000020489. input_tokens=275, output_tokens=189
|
|
15:34:19,748 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:19,749 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.109000000054948. input_tokens=149, output_tokens=80
|
|
15:34:19,995 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:19,996 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.922000000020489. input_tokens=177, output_tokens=119
|
|
15:34:20,70 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,72 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.125. input_tokens=170, output_tokens=81
|
|
15:34:20,130 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,132 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.125. input_tokens=204, output_tokens=120
|
|
15:34:20,364 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,366 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 27.390999999945052. input_tokens=437, output_tokens=459
|
|
15:34:20,387 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,388 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.515999999945052. input_tokens=146, output_tokens=54
|
|
15:34:20,405 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,407 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.061999999918044. input_tokens=206, output_tokens=157
|
|
15:34:20,479 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,481 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.85999999998603. input_tokens=143, output_tokens=85
|
|
15:34:20,498 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,499 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.969000000040978. input_tokens=165, output_tokens=122
|
|
15:34:20,607 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,608 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.922000000020489. input_tokens=788, output_tokens=500
|
|
15:34:20,702 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,705 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.437000000034459. input_tokens=166, output_tokens=104
|
|
15:34:20,993 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:20,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.9070000000065193. input_tokens=150, output_tokens=59
|
|
15:34:21,744 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:21,746 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.01500000001397. input_tokens=582, output_tokens=500
|
|
15:34:21,828 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:21,829 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.26500000001397. input_tokens=198, output_tokens=115
|
|
15:34:22,297 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:22,298 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.359000000054948. input_tokens=147, output_tokens=51
|
|
15:34:22,813 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:22,814 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.75. input_tokens=148, output_tokens=75
|
|
15:34:22,831 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:22,833 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.577999999979511. input_tokens=154, output_tokens=62
|
|
15:34:22,892 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:22,893 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.890999999945052. input_tokens=185, output_tokens=60
|
|
15:34:23,558 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:23,559 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.7349999999860302. input_tokens=144, output_tokens=24
|
|
15:34:23,624 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:23,626 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.562000000034459. input_tokens=177, output_tokens=75
|
|
15:34:23,709 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:23,710 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 26.92200000002049. input_tokens=762, output_tokens=500
|
|
15:34:23,743 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:23,745 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.5. input_tokens=164, output_tokens=65
|
|
15:34:24,370 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:24,371 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.89000000001397. input_tokens=172, output_tokens=82
|
|
15:34:24,466 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:24,468 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.969000000040978. input_tokens=161, output_tokens=92
|
|
15:34:24,715 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:24,717 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.063000000081956. input_tokens=197, output_tokens=105
|
|
15:34:25,119 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:25,120 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.734000000054948. input_tokens=198, output_tokens=111
|
|
15:34:25,262 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:25,263 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.85999999998603. input_tokens=148, output_tokens=59
|
|
15:34:25,713 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:25,715 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.109000000054948. input_tokens=166, output_tokens=128
|
|
15:34:25,718 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:25,720 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.359000000054948. input_tokens=171, output_tokens=78
|
|
15:34:25,971 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:25,973 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.969000000040978. input_tokens=208, output_tokens=143
|
|
15:34:26,25 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:26,27 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.280999999959022. input_tokens=158, output_tokens=98
|
|
15:34:26,264 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:26,265 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.968999999924563. input_tokens=160, output_tokens=55
|
|
15:34:26,315 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:26,316 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.187999999965541. input_tokens=178, output_tokens=122
|
|
15:34:26,599 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:26,602 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.891000000061467. input_tokens=233, output_tokens=156
|
|
15:34:26,734 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:26,736 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 12.90700000000652. input_tokens=493, output_tokens=251
|
|
15:34:27,389 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:27,390 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.687999999965541. input_tokens=185, output_tokens=82
|
|
15:34:27,408 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:27,411 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.51500000001397. input_tokens=170, output_tokens=114
|
|
15:34:27,593 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:27,596 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.188000000081956. input_tokens=173, output_tokens=186
|
|
15:34:28,278 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:28,279 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.3119999999180436. input_tokens=142, output_tokens=60
|
|
15:34:28,325 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:28,326 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.062000000034459. input_tokens=148, output_tokens=19
|
|
15:34:28,411 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:28,413 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.0. input_tokens=729, output_tokens=500
|
|
15:34:29,148 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:29,149 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 24.780999999959022. input_tokens=805, output_tokens=481
|
|
15:34:29,230 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:29,231 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.672000000020489. input_tokens=193, output_tokens=107
|
|
15:34:29,296 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:29,299 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.172000000020489. input_tokens=148, output_tokens=85
|
|
15:34:29,612 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:29,614 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.890999999945052. input_tokens=192, output_tokens=109
|
|
15:34:29,741 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:29,743 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.922000000020489. input_tokens=176, output_tokens=110
|
|
15:34:30,129 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,131 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.094000000040978. input_tokens=166, output_tokens=79
|
|
15:34:30,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.469000000040978. input_tokens=228, output_tokens=129
|
|
15:34:30,654 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,655 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.061999999918044. input_tokens=164, output_tokens=88
|
|
15:34:30,720 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,721 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.984000000054948. input_tokens=146, output_tokens=46
|
|
15:34:30,785 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,787 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.061999999918044. input_tokens=169, output_tokens=121
|
|
15:34:30,952 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:30,953 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.64000000001397. input_tokens=137, output_tokens=65
|
|
15:34:31,182 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:31,183 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.812999999965541. input_tokens=201, output_tokens=154
|
|
15:34:31,439 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:31,441 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 12.75. input_tokens=400, output_tokens=402
|
|
15:34:31,624 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:31,628 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.344000000040978. input_tokens=187, output_tokens=72
|
|
15:34:31,633 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:31,647 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.921999999904074. input_tokens=222, output_tokens=140
|
|
15:34:31,849 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:31,849 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.234000000054948. input_tokens=148, output_tokens=51
|
|
15:34:32,248 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.625. input_tokens=233, output_tokens=154
|
|
15:34:32,362 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,365 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.890999999945052. input_tokens=173, output_tokens=139
|
|
15:34:32,525 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,527 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.10999999998603. input_tokens=204, output_tokens=93
|
|
15:34:32,565 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,568 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.157000000006519. input_tokens=164, output_tokens=80
|
|
15:34:32,904 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,907 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6089999999385327. input_tokens=184, output_tokens=91
|
|
15:34:32,960 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:32,961 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.7179999999934807. input_tokens=178, output_tokens=111
|
|
15:34:34,89 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:34,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.344000000040978. input_tokens=227, output_tokens=314
|
|
15:34:34,410 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:34,411 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.780999999959022. input_tokens=160, output_tokens=70
|
|
15:34:34,516 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:34,517 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.218999999924563. input_tokens=183, output_tokens=75
|
|
15:34:34,703 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:34,705 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.75. input_tokens=163, output_tokens=79
|
|
15:34:34,740 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:34,746 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.562000000034459. input_tokens=198, output_tokens=100
|
|
15:34:35,238 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,239 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.875. input_tokens=179, output_tokens=83
|
|
15:34:35,290 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,292 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 11.530999999959022. input_tokens=261, output_tokens=334
|
|
15:34:35,350 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,352 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.563000000081956. input_tokens=167, output_tokens=79
|
|
15:34:35,388 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,390 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.73499999998603. input_tokens=190, output_tokens=104
|
|
15:34:35,484 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,486 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.390999999945052. input_tokens=148, output_tokens=23
|
|
15:34:35,871 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,884 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.030999999959022. input_tokens=181, output_tokens=110
|
|
15:34:35,932 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:35,934 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.218999999924563. input_tokens=150, output_tokens=110
|
|
15:34:36,114 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,120 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.0. input_tokens=203, output_tokens=189
|
|
15:34:36,137 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,138 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.5. input_tokens=165, output_tokens=97
|
|
15:34:36,139 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,143 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.812999999965541. input_tokens=248, output_tokens=236
|
|
15:34:36,270 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,272 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.35999999998603. input_tokens=173, output_tokens=103
|
|
15:34:36,805 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,806 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.656000000075437. input_tokens=186, output_tokens=181
|
|
15:34:36,868 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:36,868 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.9070000000065193. input_tokens=147, output_tokens=53
|
|
15:34:37,60 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,70 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.5. input_tokens=146, output_tokens=78
|
|
15:34:37,340 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,346 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.813000000081956. input_tokens=195, output_tokens=88
|
|
15:34:37,770 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,775 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.48499999998603. input_tokens=149, output_tokens=72
|
|
15:34:37,794 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,795 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.312000000034459. input_tokens=139, output_tokens=49
|
|
15:34:37,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,800 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.406000000075437. input_tokens=219, output_tokens=232
|
|
15:34:37,868 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:37,871 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.469000000040978. input_tokens=164, output_tokens=81
|
|
15:34:38,109 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:38,110 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.515999999945052. input_tokens=206, output_tokens=196
|
|
15:34:38,536 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:38,537 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.780999999959022. input_tokens=179, output_tokens=105
|
|
15:34:39,478 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:39,479 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.219000000040978. input_tokens=178, output_tokens=156
|
|
15:34:39,482 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:39,483 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.625. input_tokens=142, output_tokens=57
|
|
15:34:39,616 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:39,617 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.265999999945052. input_tokens=166, output_tokens=127
|
|
15:34:39,796 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:39,796 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.859000000054948. input_tokens=149, output_tokens=79
|
|
15:34:39,971 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:39,979 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.187999999965541. input_tokens=166, output_tokens=90
|
|
15:34:40,321 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:40,321 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.437999999965541. input_tokens=156, output_tokens=31
|
|
15:34:40,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:40,469 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.125. input_tokens=189, output_tokens=90
|
|
15:34:40,542 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:40,543 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.484000000054948. input_tokens=155, output_tokens=87
|
|
15:34:40,809 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:40,810 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.672000000020489. input_tokens=149, output_tokens=94
|
|
15:34:40,916 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:40,927 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.781000000075437. input_tokens=167, output_tokens=95
|
|
15:34:41,91 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:41,92 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.578000000095926. input_tokens=174, output_tokens=149
|
|
15:34:41,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:41,575 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.342999999993481. input_tokens=200, output_tokens=168
|
|
15:34:41,835 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:41,837 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.125. input_tokens=164, output_tokens=138
|
|
15:34:42,559 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:42,567 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.687999999965541. input_tokens=209, output_tokens=217
|
|
15:34:42,845 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:42,847 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.375. input_tokens=145, output_tokens=75
|
|
15:34:43,11 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,15 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.48499999998603. input_tokens=155, output_tokens=120
|
|
15:34:43,33 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,34 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.64000000001397. input_tokens=187, output_tokens=171
|
|
15:34:43,119 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,120 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.015999999945052. input_tokens=150, output_tokens=55
|
|
15:34:43,436 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,439 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.640999999945052. input_tokens=174, output_tokens=137
|
|
15:34:43,640 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,642 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.718999999924563. input_tokens=142, output_tokens=69
|
|
15:34:43,704 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,707 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.89000000001397. input_tokens=172, output_tokens=94
|
|
15:34:43,958 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:43,959 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.467999999993481. input_tokens=193, output_tokens=119
|
|
15:34:44,94 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,94 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.484000000054948. input_tokens=179, output_tokens=110
|
|
15:34:44,187 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,189 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.062999999965541. input_tokens=154, output_tokens=150
|
|
15:34:44,337 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,338 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.562000000034459. input_tokens=176, output_tokens=92
|
|
15:34:44,369 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,370 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.109000000054948. input_tokens=294, output_tokens=237
|
|
15:34:44,374 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,376 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.797000000020489. input_tokens=147, output_tokens=64
|
|
15:34:44,386 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,388 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 19.125. input_tokens=386, output_tokens=492
|
|
15:34:44,650 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,656 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.186999999918044. input_tokens=192, output_tokens=70
|
|
15:34:44,702 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,704 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 14.96799999999348. input_tokens=343, output_tokens=352
|
|
15:34:44,758 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,760 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.625. input_tokens=150, output_tokens=59
|
|
15:34:44,801 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,802 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.969000000040978. input_tokens=133, output_tokens=46
|
|
15:34:44,898 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,898 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.093999999924563. input_tokens=147, output_tokens=88
|
|
15:34:44,997 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:44,999 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.202999999979511. input_tokens=172, output_tokens=164
|
|
15:34:45,682 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:45,684 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.140999999945052. input_tokens=175, output_tokens=94
|
|
15:34:46,40 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,49 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.484000000054948. input_tokens=151, output_tokens=95
|
|
15:34:46,168 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,169 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.156000000075437. input_tokens=154, output_tokens=112
|
|
15:34:46,706 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,707 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.59299999999348. input_tokens=227, output_tokens=184
|
|
15:34:46,773 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,774 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.125. input_tokens=143, output_tokens=69
|
|
15:34:46,834 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,836 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.030999999959022. input_tokens=153, output_tokens=54
|
|
15:34:46,908 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:46,908 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.875. input_tokens=144, output_tokens=102
|
|
15:34:47,107 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,108 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.797000000020489. input_tokens=230, output_tokens=215
|
|
15:34:47,188 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,190 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.75. input_tokens=367, output_tokens=500
|
|
15:34:47,223 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,223 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.469000000040978. input_tokens=154, output_tokens=59
|
|
15:34:47,309 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,312 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.875. input_tokens=136, output_tokens=64
|
|
15:34:47,532 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,533 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.827999999979511. input_tokens=142, output_tokens=85
|
|
15:34:47,581 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,582 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.625. input_tokens=207, output_tokens=87
|
|
15:34:47,607 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,609 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.515999999945052. input_tokens=199, output_tokens=79
|
|
15:34:47,613 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:47,615 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.719000000040978. input_tokens=155, output_tokens=83
|
|
15:34:48,197 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,199 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.827999999979511. input_tokens=162, output_tokens=117
|
|
15:34:48,216 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,217 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 8.234000000054948. input_tokens=184, output_tokens=159
|
|
15:34:48,318 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,319 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.60999999998603. input_tokens=155, output_tokens=63
|
|
15:34:48,523 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,526 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.125. input_tokens=166, output_tokens=130
|
|
15:34:48,904 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,907 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.202999999979511. input_tokens=149, output_tokens=26
|
|
15:34:48,995 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:48,995 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.952999999979511. input_tokens=161, output_tokens=57
|
|
15:34:49,54 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:49,54 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.202999999979511. input_tokens=154, output_tokens=109
|
|
15:34:49,301 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:49,302 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.641000000061467. input_tokens=184, output_tokens=103
|
|
15:34:49,870 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:49,872 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.109000000054948. input_tokens=144, output_tokens=74
|
|
15:34:49,926 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:49,927 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.234000000054948. input_tokens=154, output_tokens=109
|
|
15:34:49,996 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:49,997 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.827999999979511. input_tokens=171, output_tokens=67
|
|
15:34:50,416 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:50,420 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.109000000054948. input_tokens=162, output_tokens=79
|
|
15:34:50,800 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:50,801 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.577999999979511. input_tokens=156, output_tokens=59
|
|
15:34:51,167 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:51,169 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.797000000020489. input_tokens=157, output_tokens=99
|
|
15:34:51,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:51,470 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.469000000040978. input_tokens=235, output_tokens=190
|
|
15:34:52,143 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,164 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.155999999959022. input_tokens=155, output_tokens=77
|
|
15:34:52,477 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,478 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.281000000075437. input_tokens=170, output_tokens=104
|
|
15:34:52,799 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,801 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.891000000061467. input_tokens=170, output_tokens=80
|
|
15:34:52,831 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,832 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.6559999999590218. input_tokens=155, output_tokens=24
|
|
15:34:52,926 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,934 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.827999999979511. input_tokens=169, output_tokens=93
|
|
15:34:52,994 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:52,999 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.469000000040978. input_tokens=165, output_tokens=166
|
|
15:34:53,109 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,110 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.594000000040978. input_tokens=177, output_tokens=122
|
|
15:34:53,211 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,215 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.031000000075437. input_tokens=248, output_tokens=337
|
|
15:34:53,231 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,233 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.407000000006519. input_tokens=207, output_tokens=177
|
|
15:34:53,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,289 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.967999999993481. input_tokens=225, output_tokens=162
|
|
15:34:53,548 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,550 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.5. input_tokens=144, output_tokens=80
|
|
15:34:53,971 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:53,972 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.625. input_tokens=198, output_tokens=215
|
|
15:34:54,114 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,116 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.532000000006519. input_tokens=221, output_tokens=139
|
|
15:34:54,170 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,176 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.562000000034459. input_tokens=159, output_tokens=99
|
|
15:34:54,268 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.265999999945052. input_tokens=166, output_tokens=144
|
|
15:34:54,288 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,290 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.8119999999180436. input_tokens=164, output_tokens=41
|
|
15:34:54,369 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,370 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.905999999959022. input_tokens=135, output_tokens=46
|
|
15:34:54,530 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,531 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.608999999938533. input_tokens=163, output_tokens=92
|
|
15:34:54,569 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,570 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.140999999945052. input_tokens=152, output_tokens=57
|
|
15:34:54,809 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:54,809 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.265999999945052. input_tokens=150, output_tokens=21
|
|
15:34:55,10 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,12 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.718999999924563. input_tokens=243, output_tokens=211
|
|
15:34:55,30 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,30 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.125. input_tokens=277, output_tokens=223
|
|
15:34:55,203 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,213 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.983999999938533. input_tokens=189, output_tokens=148
|
|
15:34:55,645 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,647 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.702999999979511. input_tokens=159, output_tokens=71
|
|
15:34:55,889 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,891 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.73499999998603. input_tokens=206, output_tokens=102
|
|
15:34:55,994 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:55,994 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.172000000020489. input_tokens=140, output_tokens=17
|
|
15:34:56,18 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:56,20 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.015999999945052. input_tokens=184, output_tokens=87
|
|
15:34:56,27 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:56,27 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.9209999999729916. input_tokens=169, output_tokens=90
|
|
15:34:56,287 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:56,287 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.7179999999934807. input_tokens=154, output_tokens=34
|
|
15:34:56,748 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:56,749 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.952999999979511. input_tokens=155, output_tokens=91
|
|
15:34:57,197 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,200 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.327999999979511. input_tokens=185, output_tokens=113
|
|
15:34:57,289 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,290 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.483999999938533. input_tokens=181, output_tokens=89
|
|
15:34:57,391 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,392 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.375. input_tokens=143, output_tokens=22
|
|
15:34:57,446 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,452 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.077999999979511. input_tokens=179, output_tokens=100
|
|
15:34:57,544 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,546 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.547000000020489. input_tokens=162, output_tokens=41
|
|
15:34:57,810 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:57,812 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.172000000020489. input_tokens=158, output_tokens=52
|
|
15:34:58,40 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,47 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.0160000000614673. input_tokens=145, output_tokens=72
|
|
15:34:58,70 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,73 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.890999999945052. input_tokens=153, output_tokens=83
|
|
15:34:58,418 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,421 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.812000000034459. input_tokens=184, output_tokens=201
|
|
15:34:58,438 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,439 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.202999999979511. input_tokens=194, output_tokens=172
|
|
15:34:58,505 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,506 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.219000000040978. input_tokens=171, output_tokens=69
|
|
15:34:58,777 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:58,781 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.811999999918044. input_tokens=170, output_tokens=134
|
|
15:34:58,999 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,0 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.172000000020489. input_tokens=224, output_tokens=196
|
|
15:34:59,49 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,51 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.156000000075437. input_tokens=226, output_tokens=127
|
|
15:34:59,91 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.7030000000959262. input_tokens=164, output_tokens=39
|
|
15:34:59,260 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,261 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.062999999965541. input_tokens=160, output_tokens=112
|
|
15:34:59,578 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,587 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.547000000020489. input_tokens=169, output_tokens=88
|
|
15:34:59,661 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,663 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.1089999999385327. input_tokens=154, output_tokens=57
|
|
15:34:59,790 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:34:59,794 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.266000000061467. input_tokens=167, output_tokens=85
|
|
15:35:00,27 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,30 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.811999999918044. input_tokens=190, output_tokens=100
|
|
15:35:00,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,322 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.5. input_tokens=141, output_tokens=44
|
|
15:35:00,341 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,343 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.063000000081956. input_tokens=163, output_tokens=108
|
|
15:35:00,346 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,354 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.063000000081956. input_tokens=171, output_tokens=151
|
|
15:35:00,471 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.719000000040978. input_tokens=151, output_tokens=96
|
|
15:35:00,588 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:00,590 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.484000000054948. input_tokens=165, output_tokens=136
|
|
15:35:01,302 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:01,306 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.797000000020489. input_tokens=193, output_tokens=74
|
|
15:35:02,214 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,221 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.141000000061467. input_tokens=174, output_tokens=186
|
|
15:35:02,239 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,250 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.827999999979511. input_tokens=156, output_tokens=97
|
|
15:35:02,365 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,365 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.079000000027008. input_tokens=148, output_tokens=170
|
|
15:35:02,653 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,656 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.452999999979511. input_tokens=173, output_tokens=111
|
|
15:35:02,714 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,715 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.719000000040978. input_tokens=192, output_tokens=79
|
|
15:35:02,933 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:02,935 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.672000000020489. input_tokens=170, output_tokens=58
|
|
15:35:03,59 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,68 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,69 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.60999999998603. input_tokens=150, output_tokens=61
|
|
15:35:03,141 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,143 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.093999999924563. input_tokens=159, output_tokens=78
|
|
15:35:03,148 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.35999999998603. input_tokens=188, output_tokens=84
|
|
15:35:03,268 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,269 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.0. input_tokens=170, output_tokens=183
|
|
15:35:03,391 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,393 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 15.187999999965541. input_tokens=336, output_tokens=500
|
|
15:35:03,505 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.405999999959022. input_tokens=194, output_tokens=107
|
|
15:35:03,688 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,691 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.218999999924563. input_tokens=170, output_tokens=80
|
|
15:35:03,780 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,781 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.9839999999385327. input_tokens=176, output_tokens=108
|
|
15:35:03,963 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:03,964 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.656000000075437. input_tokens=191, output_tokens=95
|
|
15:35:04,30 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:04,30 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.375. input_tokens=254, output_tokens=124
|
|
15:35:04,445 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:04,448 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.218999999924563. input_tokens=199, output_tokens=87
|
|
15:35:04,454 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:04,455 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.01500000001397. input_tokens=197, output_tokens=74
|
|
15:35:04,471 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:04,472 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.422000000020489. input_tokens=375, output_tokens=143
|
|
15:35:04,668 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:04,669 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.375. input_tokens=182, output_tokens=75
|
|
15:35:05,159 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,160 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.14000000001397. input_tokens=165, output_tokens=151
|
|
15:35:05,177 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,178 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.141000000061467. input_tokens=192, output_tokens=92
|
|
15:35:05,425 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,429 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.844000000040978. input_tokens=197, output_tokens=116
|
|
15:35:05,445 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,450 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.797000000020489. input_tokens=195, output_tokens=74
|
|
15:35:05,556 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,557 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.202999999979511. input_tokens=189, output_tokens=90
|
|
15:35:05,836 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:05,837 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.562000000034459. input_tokens=191, output_tokens=73
|
|
15:35:06,27 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,28 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6709999999729916. input_tokens=194, output_tokens=76
|
|
15:35:06,65 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,67 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.468999999924563. input_tokens=163, output_tokens=84
|
|
15:35:06,332 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,335 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.187000000034459. input_tokens=194, output_tokens=59
|
|
15:35:06,352 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,354 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.0. input_tokens=222, output_tokens=121
|
|
15:35:06,480 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,482 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.094000000040978. input_tokens=186, output_tokens=59
|
|
15:35:06,672 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,674 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.6410000000614673. input_tokens=169, output_tokens=65
|
|
15:35:06,872 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.155999999959022. input_tokens=201, output_tokens=122
|
|
15:35:06,908 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,908 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.9679999999934807. input_tokens=175, output_tokens=83
|
|
15:35:06,991 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:06,993 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.73499999998603. input_tokens=182, output_tokens=99
|
|
15:35:07,116 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:07,117 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.422000000020489. input_tokens=190, output_tokens=105
|
|
15:35:07,156 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:07,157 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.702999999979511. input_tokens=187, output_tokens=57
|
|
15:35:07,371 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:07,373 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.905999999959022. input_tokens=205, output_tokens=59
|
|
15:35:07,427 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:07,428 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6410000000614673. input_tokens=168, output_tokens=79
|
|
15:35:07,840 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:07,841 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.781000000075437. input_tokens=193, output_tokens=89
|
|
15:35:08,8 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:08,9 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.577999999979511. input_tokens=183, output_tokens=58
|
|
15:35:08,484 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:08,486 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.6570000000065193. input_tokens=161, output_tokens=75
|
|
15:35:08,908 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:08,910 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.7339999999385327. input_tokens=174, output_tokens=56
|
|
15:35:08,926 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:08,927 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.562000000034459. input_tokens=159, output_tokens=72
|
|
15:35:09,0 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,1 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.562000000034459. input_tokens=191, output_tokens=86
|
|
15:35:09,381 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,382 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.922000000020489. input_tokens=190, output_tokens=79
|
|
15:35:09,572 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,574 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.905999999959022. input_tokens=194, output_tokens=86
|
|
15:35:09,774 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,775 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.343999999924563. input_tokens=165, output_tokens=93
|
|
15:35:09,809 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,810 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 1.9689999999245629. input_tokens=158, output_tokens=51
|
|
15:35:09,869 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,871 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.375. input_tokens=194, output_tokens=129
|
|
15:35:09,987 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:09,989 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.9540000000270084. input_tokens=150, output_tokens=70
|
|
15:35:10,136 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,137 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.593999999924563. input_tokens=183, output_tokens=78
|
|
15:35:10,288 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,293 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.234000000054948. input_tokens=174, output_tokens=144
|
|
15:35:10,397 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,399 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.25. input_tokens=235, output_tokens=181
|
|
15:35:10,586 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,586 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5929999999934807. input_tokens=213, output_tokens=128
|
|
15:35:10,744 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,744 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.625. input_tokens=167, output_tokens=58
|
|
15:35:10,765 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:10,767 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.765999999945052. input_tokens=165, output_tokens=72
|
|
15:35:11,107 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,109 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.140999999945052. input_tokens=165, output_tokens=101
|
|
15:35:11,508 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,509 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.172000000020489. input_tokens=178, output_tokens=98
|
|
15:35:11,595 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,596 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.688000000081956. input_tokens=191, output_tokens=119
|
|
15:35:11,737 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,741 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.85999999998603. input_tokens=188, output_tokens=129
|
|
15:35:11,891 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,892 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.405999999959022. input_tokens=180, output_tokens=97
|
|
15:35:11,970 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:11,971 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.594000000040978. input_tokens=175, output_tokens=78
|
|
15:35:12,81 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,82 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.077999999979511. input_tokens=171, output_tokens=78
|
|
15:35:12,157 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,158 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.0. input_tokens=255, output_tokens=196
|
|
15:35:12,661 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,663 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.077999999979511. input_tokens=162, output_tokens=84
|
|
15:35:12,760 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,762 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.952999999979511. input_tokens=175, output_tokens=83
|
|
15:35:12,839 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,842 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.0780000000959262. input_tokens=162, output_tokens=84
|
|
15:35:12,921 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:12,922 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.937000000034459. input_tokens=171, output_tokens=47
|
|
15:35:13,357 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:13,359 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.437999999965541. input_tokens=185, output_tokens=139
|
|
15:35:13,574 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:13,575 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.280999999959022. input_tokens=171, output_tokens=91
|
|
15:35:13,619 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:13,622 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.14000000001397. input_tokens=271, output_tokens=144
|
|
15:35:13,641 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:13,642 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.48499999998603. input_tokens=223, output_tokens=131
|
|
15:35:13,688 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:13,690 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.187999999965541. input_tokens=196, output_tokens=73
|
|
15:35:14,17 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:14,19 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.10999999998603. input_tokens=168, output_tokens=81
|
|
15:35:14,195 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:14,197 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.60999999998603. input_tokens=205, output_tokens=101
|
|
15:35:14,351 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:14,352 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.969000000040978. input_tokens=215, output_tokens=154
|
|
15:35:14,514 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:14,516 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.640999999945052. input_tokens=188, output_tokens=73
|
|
15:35:14,996 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:14,997 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.609000000054948. input_tokens=175, output_tokens=104
|
|
15:35:15,317 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,318 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.23499999998603. input_tokens=203, output_tokens=81
|
|
15:35:15,407 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,408 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.295999999972992. input_tokens=175, output_tokens=155
|
|
15:35:15,444 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,446 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.297000000020489. input_tokens=215, output_tokens=93
|
|
15:35:15,555 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,557 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.952999999979511. input_tokens=169, output_tokens=117
|
|
15:35:15,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,571 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.4070000000065193. input_tokens=173, output_tokens=89
|
|
15:35:15,667 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,667 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.906000000075437. input_tokens=181, output_tokens=78
|
|
15:35:15,828 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,830 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.905999999959022. input_tokens=166, output_tokens=57
|
|
15:35:15,842 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:15,844 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.109000000054948. input_tokens=166, output_tokens=82
|
|
15:35:16,296 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:16,299 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.6410000000614673. input_tokens=189, output_tokens=81
|
|
15:35:16,601 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:16,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.406000000075437. input_tokens=179, output_tokens=66
|
|
15:35:16,660 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:16,663 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.030999999959022. input_tokens=180, output_tokens=44
|
|
15:35:16,728 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:16,737 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.375. input_tokens=165, output_tokens=98
|
|
15:35:17,93 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,100 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.25. input_tokens=200, output_tokens=132
|
|
15:35:17,124 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,126 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 10.452999999979511. input_tokens=215, output_tokens=220
|
|
15:35:17,284 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,286 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 3.5929999999934807. input_tokens=203, output_tokens=99
|
|
15:35:17,412 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,413 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.64000000001397. input_tokens=184, output_tokens=77
|
|
15:35:17,624 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,627 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 2.312000000034459. input_tokens=196, output_tokens=66
|
|
15:35:17,819 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:17,820 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.23499999998603. input_tokens=174, output_tokens=46
|
|
15:35:18,37 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:18,39 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.295999999972992. input_tokens=253, output_tokens=118
|
|
15:35:18,170 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:18,291 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:18,293 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 6.327999999979511. input_tokens=230, output_tokens=115
|
|
15:35:18,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:18,686 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.047000000020489. input_tokens=154, output_tokens=51
|
|
15:35:19,205 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:19,207 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.858999999938533. input_tokens=173, output_tokens=77
|
|
15:35:19,582 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:19,584 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 5.562000000034459. input_tokens=178, output_tokens=109
|
|
15:35:19,720 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:19,721 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.828000000095926. input_tokens=293, output_tokens=131
|
|
15:35:19,970 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 4.969000000040978. input_tokens=190, output_tokens=83
|
|
15:35:22,245 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:22,248 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 7.734000000054948. input_tokens=204, output_tokens=182
|
|
15:35:24,503 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:35:24,505 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "summarize" with 0 retries took 9.094000000040978. input_tokens=206, output_tokens=196
|
|
15:35:24,623 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
15:35:24,632 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_summarized_entities.parquet
|
|
15:35:24,777 graphrag.index.run.workflow INFO dependencies for create_base_entity_graph: ['create_summarized_entities']
|
|
15:35:24,778 graphrag.utils.storage INFO read table from storage: create_summarized_entities.parquet
|
|
15:35:24,803 datashaper.workflow.workflow INFO executing verb cluster_graph
|
|
15:35:25,951 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
15:35:25,973 datashaper.workflow.workflow INFO executing verb embed_graph
|
|
15:35:26,73 root INFO Starting preprocessing of transition probabilities on graph with 882 nodes and 1287 edges
|
|
15:35:26,73 root INFO Starting at time 1751528126.0739422
|
|
15:35:26,73 root INFO Beginning preprocessing of transition probabilities for 882 vertices
|
|
15:35:26,73 root INFO Completed 1 / 882 vertices
|
|
15:35:26,75 root INFO Completed 89 / 882 vertices
|
|
15:35:26,76 root INFO Completed 177 / 882 vertices
|
|
15:35:26,76 root INFO Completed 265 / 882 vertices
|
|
15:35:26,77 root INFO Completed 353 / 882 vertices
|
|
15:35:26,77 root INFO Completed 441 / 882 vertices
|
|
15:35:26,78 root INFO Completed 529 / 882 vertices
|
|
15:35:26,79 root INFO Completed 617 / 882 vertices
|
|
15:35:26,79 root INFO Completed 705 / 882 vertices
|
|
15:35:26,79 root INFO Completed 793 / 882 vertices
|
|
15:35:26,79 root INFO Completed 881 / 882 vertices
|
|
15:35:26,79 root INFO Completed preprocessing of transition probabilities for vertices
|
|
15:35:26,80 root INFO Beginning preprocessing of transition probabilities for 1287 edges
|
|
15:35:26,80 root INFO Completed 1 / 1287 edges
|
|
15:35:26,89 root INFO Completed 129 / 1287 edges
|
|
15:35:26,101 root INFO Completed 257 / 1287 edges
|
|
15:35:26,107 root INFO Completed 385 / 1287 edges
|
|
15:35:26,117 root INFO Completed 513 / 1287 edges
|
|
15:35:26,126 root INFO Completed 641 / 1287 edges
|
|
15:35:26,132 root INFO Completed 769 / 1287 edges
|
|
15:35:26,137 root INFO Completed 897 / 1287 edges
|
|
15:35:26,140 root INFO Completed 1025 / 1287 edges
|
|
15:35:26,143 root INFO Completed 1153 / 1287 edges
|
|
15:35:26,145 root INFO Completed 1281 / 1287 edges
|
|
15:35:26,145 root INFO Completed preprocessing of transition probabilities for edges
|
|
15:35:26,145 root INFO Simulating walks on graph at time 1751528126.1456225
|
|
15:35:26,147 root INFO Walk iteration: 1/10
|
|
15:35:26,199 root INFO Walk iteration: 2/10
|
|
15:35:26,247 root INFO Walk iteration: 3/10
|
|
15:35:26,296 root INFO Walk iteration: 4/10
|
|
15:35:26,346 root INFO Walk iteration: 5/10
|
|
15:35:26,394 root INFO Walk iteration: 6/10
|
|
15:35:26,443 root INFO Walk iteration: 7/10
|
|
15:35:26,491 root INFO Walk iteration: 8/10
|
|
15:35:26,551 root INFO Walk iteration: 9/10
|
|
15:35:26,606 root INFO Walk iteration: 10/10
|
|
15:35:26,659 root INFO Learning embeddings at time 1751528126.6590097
|
|
15:35:26,669 gensim.models.word2vec INFO collecting all words and their counts
|
|
15:35:26,669 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
15:35:26,680 gensim.models.word2vec INFO collected 882 word types from a corpus of 158560 raw words and 8820 sentences
|
|
15:35:26,680 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
15:35:26,682 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 882 unique words (100.00% of original 882, drops 0)', 'datetime': '2025-07-03T15:35:26.682156', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:26,682 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 158560 word corpus (100.00% of original 158560, drops 0)', 'datetime': '2025-07-03T15:35:26.682156', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:26,685 gensim.models.word2vec INFO deleting the raw counts dictionary of 882 items
|
|
15:35:26,686 gensim.models.word2vec INFO sample=0.001 downsamples 67 most-common words
|
|
15:35:26,686 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 122971.75848756742 word corpus (77.6%% of prior 158560)', 'datetime': '2025-07-03T15:35:26.686636', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:26,691 gensim.models.word2vec INFO estimated required memory for 882 words and 1536 dimensions: 11279016 bytes
|
|
15:35:26,691 gensim.models.word2vec INFO resetting layer weights
|
|
15:35:26,697 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-07-03T15:35:26.697808', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
15:35:26,697 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 882 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-07-03T15:35:26.697808', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:26,917 gensim.models.word2vec INFO EPOCH 0: training on 158560 raw words (122953 effective words) took 0.2s, 586390 effective words/s
|
|
15:35:27,142 gensim.models.word2vec INFO EPOCH 1: training on 158560 raw words (123026 effective words) took 0.2s, 566339 effective words/s
|
|
15:35:27,362 gensim.models.word2vec INFO EPOCH 2: training on 158560 raw words (122888 effective words) took 0.2s, 578101 effective words/s
|
|
15:35:27,362 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 475680 raw words (368867 effective words) took 0.7s, 555225 effective words/s', 'datetime': '2025-07-03T15:35:27.362687', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:27,362 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=882, vector_size=1536, alpha=0.025>', 'datetime': '2025-07-03T15:35:27.362687', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
15:35:27,363 root INFO Completed. Ending time is 1751528127.3634343 Elapsed time is -1.289492130279541
|
|
15:35:27,627 root INFO Starting preprocessing of transition probabilities on graph with 882 nodes and 1287 edges
|
|
15:35:27,627 root INFO Starting at time 1751528127.6270275
|
|
15:35:27,627 root INFO Beginning preprocessing of transition probabilities for 882 vertices
|
|
15:35:27,627 root INFO Completed 1 / 882 vertices
|
|
15:35:27,628 root INFO Completed 89 / 882 vertices
|
|
15:35:27,629 root INFO Completed 177 / 882 vertices
|
|
15:35:27,629 root INFO Completed 265 / 882 vertices
|
|
15:35:27,630 root INFO Completed 353 / 882 vertices
|
|
15:35:27,631 root INFO Completed 441 / 882 vertices
|
|
15:35:27,631 root INFO Completed 529 / 882 vertices
|
|
15:35:27,632 root INFO Completed 617 / 882 vertices
|
|
15:35:27,632 root INFO Completed 705 / 882 vertices
|
|
15:35:27,632 root INFO Completed 793 / 882 vertices
|
|
15:35:27,632 root INFO Completed 881 / 882 vertices
|
|
15:35:27,632 root INFO Completed preprocessing of transition probabilities for vertices
|
|
15:35:27,633 root INFO Beginning preprocessing of transition probabilities for 1287 edges
|
|
15:35:27,633 root INFO Completed 1 / 1287 edges
|
|
15:35:27,642 root INFO Completed 129 / 1287 edges
|
|
15:35:27,654 root INFO Completed 257 / 1287 edges
|
|
15:35:27,660 root INFO Completed 385 / 1287 edges
|
|
15:35:27,669 root INFO Completed 513 / 1287 edges
|
|
15:35:27,678 root INFO Completed 641 / 1287 edges
|
|
15:35:27,684 root INFO Completed 769 / 1287 edges
|
|
15:35:27,690 root INFO Completed 897 / 1287 edges
|
|
15:35:27,692 root INFO Completed 1025 / 1287 edges
|
|
15:35:27,695 root INFO Completed 1153 / 1287 edges
|
|
15:35:27,698 root INFO Completed 1281 / 1287 edges
|
|
15:35:27,698 root INFO Completed preprocessing of transition probabilities for edges
|
|
15:35:27,698 root INFO Simulating walks on graph at time 1751528127.698223
|
|
15:35:27,698 root INFO Walk iteration: 1/10
|
|
15:35:27,759 root INFO Walk iteration: 2/10
|
|
15:35:27,812 root INFO Walk iteration: 3/10
|
|
15:35:27,865 root INFO Walk iteration: 4/10
|
|
15:35:27,917 root INFO Walk iteration: 5/10
|
|
15:35:27,969 root INFO Walk iteration: 6/10
|
|
15:35:28,23 root INFO Walk iteration: 7/10
|
|
15:35:28,76 root INFO Walk iteration: 8/10
|
|
15:35:28,126 root INFO Walk iteration: 9/10
|
|
15:35:28,176 root INFO Walk iteration: 10/10
|
|
15:35:28,225 root INFO Learning embeddings at time 1751528128.2253993
|
|
15:35:28,235 gensim.models.word2vec INFO collecting all words and their counts
|
|
15:35:28,235 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
15:35:28,246 gensim.models.word2vec INFO collected 882 word types from a corpus of 158560 raw words and 8820 sentences
|
|
15:35:28,246 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
15:35:28,247 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 882 unique words (100.00% of original 882, drops 0)', 'datetime': '2025-07-03T15:35:28.247621', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:28,248 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 158560 word corpus (100.00% of original 158560, drops 0)', 'datetime': '2025-07-03T15:35:28.248377', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:28,251 gensim.models.word2vec INFO deleting the raw counts dictionary of 882 items
|
|
15:35:28,251 gensim.models.word2vec INFO sample=0.001 downsamples 67 most-common words
|
|
15:35:28,251 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 122971.75848756742 word corpus (77.6%% of prior 158560)', 'datetime': '2025-07-03T15:35:28.251346', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:28,255 gensim.models.word2vec INFO estimated required memory for 882 words and 1536 dimensions: 11279016 bytes
|
|
15:35:28,256 gensim.models.word2vec INFO resetting layer weights
|
|
15:35:28,261 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-07-03T15:35:28.261605', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
15:35:28,261 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 882 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-07-03T15:35:28.261605', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:28,492 gensim.models.word2vec INFO EPOCH 0: training on 158560 raw words (123236 effective words) took 0.2s, 552483 effective words/s
|
|
15:35:28,712 gensim.models.word2vec INFO EPOCH 1: training on 158560 raw words (122765 effective words) took 0.2s, 577928 effective words/s
|
|
15:35:28,934 gensim.models.word2vec INFO EPOCH 2: training on 158560 raw words (123133 effective words) took 0.2s, 574347 effective words/s
|
|
15:35:28,934 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 475680 raw words (369134 effective words) took 0.7s, 548368 effective words/s', 'datetime': '2025-07-03T15:35:28.934726', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:28,934 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=882, vector_size=1536, alpha=0.025>', 'datetime': '2025-07-03T15:35:28.934726', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
15:35:28,935 root INFO Completed. Ending time is 1751528128.9354737 Elapsed time is -1.3084461688995361
|
|
15:35:29,76 root INFO Starting preprocessing of transition probabilities on graph with 882 nodes and 1287 edges
|
|
15:35:29,76 root INFO Starting at time 1751528129.0765953
|
|
15:35:29,76 root INFO Beginning preprocessing of transition probabilities for 882 vertices
|
|
15:35:29,76 root INFO Completed 1 / 882 vertices
|
|
15:35:29,78 root INFO Completed 89 / 882 vertices
|
|
15:35:29,78 root INFO Completed 177 / 882 vertices
|
|
15:35:29,79 root INFO Completed 265 / 882 vertices
|
|
15:35:29,80 root INFO Completed 353 / 882 vertices
|
|
15:35:29,80 root INFO Completed 441 / 882 vertices
|
|
15:35:29,81 root INFO Completed 529 / 882 vertices
|
|
15:35:29,81 root INFO Completed 617 / 882 vertices
|
|
15:35:29,81 root INFO Completed 705 / 882 vertices
|
|
15:35:29,82 root INFO Completed 793 / 882 vertices
|
|
15:35:29,82 root INFO Completed 881 / 882 vertices
|
|
15:35:29,82 root INFO Completed preprocessing of transition probabilities for vertices
|
|
15:35:29,82 root INFO Beginning preprocessing of transition probabilities for 1287 edges
|
|
15:35:29,82 root INFO Completed 1 / 1287 edges
|
|
15:35:29,91 root INFO Completed 129 / 1287 edges
|
|
15:35:29,108 root INFO Completed 257 / 1287 edges
|
|
15:35:29,119 root INFO Completed 385 / 1287 edges
|
|
15:35:29,129 root INFO Completed 513 / 1287 edges
|
|
15:35:29,139 root INFO Completed 641 / 1287 edges
|
|
15:35:29,145 root INFO Completed 769 / 1287 edges
|
|
15:35:29,152 root INFO Completed 897 / 1287 edges
|
|
15:35:29,155 root INFO Completed 1025 / 1287 edges
|
|
15:35:29,157 root INFO Completed 1153 / 1287 edges
|
|
15:35:29,159 root INFO Completed 1281 / 1287 edges
|
|
15:35:29,160 root INFO Completed preprocessing of transition probabilities for edges
|
|
15:35:29,160 root INFO Simulating walks on graph at time 1751528129.1602442
|
|
15:35:29,160 root INFO Walk iteration: 1/10
|
|
15:35:29,213 root INFO Walk iteration: 2/10
|
|
15:35:29,265 root INFO Walk iteration: 3/10
|
|
15:35:29,318 root INFO Walk iteration: 4/10
|
|
15:35:29,373 root INFO Walk iteration: 5/10
|
|
15:35:29,426 root INFO Walk iteration: 6/10
|
|
15:35:29,479 root INFO Walk iteration: 7/10
|
|
15:35:29,531 root INFO Walk iteration: 8/10
|
|
15:35:29,581 root INFO Walk iteration: 9/10
|
|
15:35:29,631 root INFO Walk iteration: 10/10
|
|
15:35:29,679 root INFO Learning embeddings at time 1751528129.679928
|
|
15:35:29,815 gensim.models.word2vec INFO collecting all words and their counts
|
|
15:35:29,815 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
15:35:29,826 gensim.models.word2vec INFO collected 882 word types from a corpus of 158560 raw words and 8820 sentences
|
|
15:35:29,826 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
15:35:29,828 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 882 unique words (100.00% of original 882, drops 0)', 'datetime': '2025-07-03T15:35:29.828961', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:29,828 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 158560 word corpus (100.00% of original 158560, drops 0)', 'datetime': '2025-07-03T15:35:29.828961', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:29,831 gensim.models.word2vec INFO deleting the raw counts dictionary of 882 items
|
|
15:35:29,831 gensim.models.word2vec INFO sample=0.001 downsamples 67 most-common words
|
|
15:35:29,831 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 122971.75848756742 word corpus (77.6%% of prior 158560)', 'datetime': '2025-07-03T15:35:29.831977', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:29,837 gensim.models.word2vec INFO estimated required memory for 882 words and 1536 dimensions: 11279016 bytes
|
|
15:35:29,837 gensim.models.word2vec INFO resetting layer weights
|
|
15:35:29,842 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-07-03T15:35:29.842226', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
15:35:29,842 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 882 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-07-03T15:35:29.842226', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:30,67 gensim.models.word2vec INFO EPOCH 0: training on 158560 raw words (122852 effective words) took 0.2s, 566243 effective words/s
|
|
15:35:30,292 gensim.models.word2vec INFO EPOCH 1: training on 158560 raw words (122802 effective words) took 0.2s, 564605 effective words/s
|
|
15:35:30,525 gensim.models.word2vec INFO EPOCH 2: training on 158560 raw words (122819 effective words) took 0.2s, 548151 effective words/s
|
|
15:35:30,525 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 475680 raw words (368473 effective words) took 0.7s, 539510 effective words/s', 'datetime': '2025-07-03T15:35:30.525464', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:30,525 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=882, vector_size=1536, alpha=0.025>', 'datetime': '2025-07-03T15:35:30.525464', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
15:35:30,526 root INFO Completed. Ending time is 1751528130.5262103 Elapsed time is -1.4496150016784668
|
|
15:35:30,667 root INFO Starting preprocessing of transition probabilities on graph with 882 nodes and 1287 edges
|
|
15:35:30,667 root INFO Starting at time 1751528130.6679754
|
|
15:35:30,667 root INFO Beginning preprocessing of transition probabilities for 882 vertices
|
|
15:35:30,668 root INFO Completed 1 / 882 vertices
|
|
15:35:30,669 root INFO Completed 89 / 882 vertices
|
|
15:35:30,670 root INFO Completed 177 / 882 vertices
|
|
15:35:30,670 root INFO Completed 265 / 882 vertices
|
|
15:35:30,671 root INFO Completed 353 / 882 vertices
|
|
15:35:30,672 root INFO Completed 441 / 882 vertices
|
|
15:35:30,672 root INFO Completed 529 / 882 vertices
|
|
15:35:30,673 root INFO Completed 617 / 882 vertices
|
|
15:35:30,673 root INFO Completed 705 / 882 vertices
|
|
15:35:30,673 root INFO Completed 793 / 882 vertices
|
|
15:35:30,673 root INFO Completed 881 / 882 vertices
|
|
15:35:30,673 root INFO Completed preprocessing of transition probabilities for vertices
|
|
15:35:30,674 root INFO Beginning preprocessing of transition probabilities for 1287 edges
|
|
15:35:30,674 root INFO Completed 1 / 1287 edges
|
|
15:35:30,682 root INFO Completed 129 / 1287 edges
|
|
15:35:30,694 root INFO Completed 257 / 1287 edges
|
|
15:35:30,701 root INFO Completed 385 / 1287 edges
|
|
15:35:30,711 root INFO Completed 513 / 1287 edges
|
|
15:35:30,719 root INFO Completed 641 / 1287 edges
|
|
15:35:30,725 root INFO Completed 769 / 1287 edges
|
|
15:35:30,731 root INFO Completed 897 / 1287 edges
|
|
15:35:30,734 root INFO Completed 1025 / 1287 edges
|
|
15:35:30,736 root INFO Completed 1153 / 1287 edges
|
|
15:35:30,738 root INFO Completed 1281 / 1287 edges
|
|
15:35:30,739 root INFO Completed preprocessing of transition probabilities for edges
|
|
15:35:30,739 root INFO Simulating walks on graph at time 1751528130.739676
|
|
15:35:30,739 root INFO Walk iteration: 1/10
|
|
15:35:30,788 root INFO Walk iteration: 2/10
|
|
15:35:30,838 root INFO Walk iteration: 3/10
|
|
15:35:30,887 root INFO Walk iteration: 4/10
|
|
15:35:30,935 root INFO Walk iteration: 5/10
|
|
15:35:30,983 root INFO Walk iteration: 6/10
|
|
15:35:31,32 root INFO Walk iteration: 7/10
|
|
15:35:31,82 root INFO Walk iteration: 8/10
|
|
15:35:31,130 root INFO Walk iteration: 9/10
|
|
15:35:31,179 root INFO Walk iteration: 10/10
|
|
15:35:31,228 root INFO Learning embeddings at time 1751528131.228094
|
|
15:35:31,237 gensim.models.word2vec INFO collecting all words and their counts
|
|
15:35:31,237 gensim.models.word2vec INFO PROGRESS: at sentence #0, processed 0 words, keeping 0 word types
|
|
15:35:31,248 gensim.models.word2vec INFO collected 882 word types from a corpus of 158560 raw words and 8820 sentences
|
|
15:35:31,248 gensim.models.word2vec INFO Creating a fresh vocabulary
|
|
15:35:31,250 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 retains 882 unique words (100.00% of original 882, drops 0)', 'datetime': '2025-07-03T15:35:31.250499', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:31,250 gensim.utils INFO Word2Vec lifecycle event {'msg': 'effective_min_count=0 leaves 158560 word corpus (100.00% of original 158560, drops 0)', 'datetime': '2025-07-03T15:35:31.250499', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:31,254 gensim.models.word2vec INFO deleting the raw counts dictionary of 882 items
|
|
15:35:31,254 gensim.models.word2vec INFO sample=0.001 downsamples 67 most-common words
|
|
15:35:31,254 gensim.utils INFO Word2Vec lifecycle event {'msg': 'downsampling leaves estimated 122971.75848756742 word corpus (77.6%% of prior 158560)', 'datetime': '2025-07-03T15:35:31.254227', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'prepare_vocab'}
|
|
15:35:31,259 gensim.models.word2vec INFO estimated required memory for 882 words and 1536 dimensions: 11279016 bytes
|
|
15:35:31,259 gensim.models.word2vec INFO resetting layer weights
|
|
15:35:31,262 gensim.utils INFO Word2Vec lifecycle event {'update': False, 'trim_rule': 'None', 'datetime': '2025-07-03T15:35:31.262981', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'build_vocab'}
|
|
15:35:31,262 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training model with 8 workers on 882 vocabulary and 1536 features, using sg=1 hs=0 sample=0.001 negative=5 window=2 shrink_windows=True', 'datetime': '2025-07-03T15:35:31.262981', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:31,486 gensim.models.word2vec INFO EPOCH 0: training on 158560 raw words (122932 effective words) took 0.2s, 570376 effective words/s
|
|
15:35:31,713 gensim.models.word2vec INFO EPOCH 1: training on 158560 raw words (122892 effective words) took 0.2s, 559742 effective words/s
|
|
15:35:31,937 gensim.models.word2vec INFO EPOCH 2: training on 158560 raw words (122795 effective words) took 0.2s, 568489 effective words/s
|
|
15:35:31,937 gensim.utils INFO Word2Vec lifecycle event {'msg': 'training on 475680 raw words (368619 effective words) took 0.7s, 546145 effective words/s', 'datetime': '2025-07-03T15:35:31.937937', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'train'}
|
|
15:35:31,937 gensim.utils INFO Word2Vec lifecycle event {'params': 'Word2Vec<vocab=882, vector_size=1536, alpha=0.025>', 'datetime': '2025-07-03T15:35:31.937937', 'gensim': '4.3.3', 'python': '3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]', 'platform': 'Windows-10-10.0.19045-SP0', 'event': 'created'}
|
|
15:35:31,938 root INFO Completed. Ending time is 1751528131.9386847 Elapsed time is -1.2707092761993408
|
|
15:35:31,992 datashaper.workflow.workflow INFO executing verb snapshot_rows
|
|
15:35:32,14 datashaper.workflow.workflow INFO executing verb select
|
|
15:35:32,135 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_entity_graph.parquet
|
|
15:35:33,485 graphrag.index.run.workflow INFO dependencies for create_final_entities: ['create_base_entity_graph']
|
|
15:35:33,486 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
15:35:33,738 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
15:35:34,214 datashaper.workflow.workflow INFO executing verb rename
|
|
15:35:34,226 datashaper.workflow.workflow INFO executing verb select
|
|
15:35:34,240 datashaper.workflow.workflow INFO executing verb dedupe
|
|
15:35:34,254 datashaper.workflow.workflow INFO executing verb rename
|
|
15:35:34,267 datashaper.workflow.workflow INFO executing verb filter
|
|
15:35:34,312 datashaper.workflow.workflow INFO executing verb text_split
|
|
15:35:34,349 datashaper.workflow.workflow INFO executing verb drop
|
|
15:35:34,363 datashaper.workflow.workflow INFO executing verb merge
|
|
15:35:34,726 datashaper.workflow.workflow INFO executing verb text_embed
|
|
15:35:34,727 graphrag.llm.openai.create_openai_client INFO Creating OpenAI client base_url=None
|
|
15:35:34,736 graphrag.index.llm.load_llm INFO create TPM/RPM limiter for text-embedding-3-small: TPM=0, RPM=0
|
|
15:35:34,736 graphrag.index.llm.load_llm INFO create concurrency limiter for text-embedding-3-small: 25
|
|
15:35:34,845 graphrag.index.verbs.text.embed.strategies.openai INFO embedding 1612 inputs via 1612 snippets using 101 batches. max_batch_size=16, max_tokens=8191
|
|
15:35:35,474 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,553 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,591 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,676 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,711 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,744 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,771 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,773 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,774 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,789 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,795 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,797 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,822 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,836 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,857 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9849999999860302. input_tokens=544, output_tokens=0
|
|
15:35:35,875 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,903 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0459999999729916. input_tokens=1501, output_tokens=0
|
|
15:35:35,920 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,952 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,957 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:35,961 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.077999999979511. input_tokens=1106, output_tokens=0
|
|
15:35:36,3 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.125. input_tokens=866, output_tokens=0
|
|
15:35:36,26 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,34 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.1559999999590218. input_tokens=931, output_tokens=0
|
|
15:35:36,53 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,161 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,219 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3440000000409782. input_tokens=864, output_tokens=0
|
|
15:35:36,243 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.375. input_tokens=977, output_tokens=0
|
|
15:35:36,262 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4059999999590218. input_tokens=1151, output_tokens=0
|
|
15:35:36,301 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.422000000020489. input_tokens=754, output_tokens=0
|
|
15:35:36,319 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,323 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.452999999979511. input_tokens=577, output_tokens=0
|
|
15:35:36,352 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,352 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,360 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5. input_tokens=3148, output_tokens=0
|
|
15:35:36,377 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5. input_tokens=477, output_tokens=0
|
|
15:35:36,416 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,423 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.547000000020489. input_tokens=318, output_tokens=0
|
|
15:35:36,441 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.577999999979511. input_tokens=387, output_tokens=0
|
|
15:35:36,464 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5940000000409782. input_tokens=1322, output_tokens=0
|
|
15:35:36,482 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6099999999860302. input_tokens=555, output_tokens=0
|
|
15:35:36,500 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,535 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.6709999999729916. input_tokens=3168, output_tokens=0
|
|
15:35:36,556 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.672000000020489. input_tokens=937, output_tokens=0
|
|
15:35:36,579 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7179999999934807. input_tokens=2599, output_tokens=0
|
|
15:35:36,604 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.734000000054948. input_tokens=988, output_tokens=0
|
|
15:35:36,630 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,659 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.7809999999590218. input_tokens=1253, output_tokens=0
|
|
15:35:36,677 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.797000000020489. input_tokens=330, output_tokens=0
|
|
15:35:36,692 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,746 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.8900000000139698. input_tokens=2819, output_tokens=0
|
|
15:35:36,762 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,766 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8439999999245629. input_tokens=828, output_tokens=0
|
|
15:35:36,788 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.9059999999590218. input_tokens=270, output_tokens=0
|
|
15:35:36,822 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5629999999655411. input_tokens=363, output_tokens=0
|
|
15:35:36,844 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.9530000000959262. input_tokens=484, output_tokens=0
|
|
15:35:36,862 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,863 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:36,939 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.062999999965541. input_tokens=1149, output_tokens=0
|
|
15:35:36,956 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,39 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,48 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,50 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,51 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,82 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,84 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,86 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,90 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,92 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.047000000020489. input_tokens=676, output_tokens=0
|
|
15:35:37,110 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,131 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,138 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,153 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9209999999729916. input_tokens=663, output_tokens=0
|
|
15:35:37,168 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,216 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,219 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,253 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,295 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5310000000754371. input_tokens=634, output_tokens=0
|
|
15:35:37,312 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,315 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.297000000020489. input_tokens=893, output_tokens=0
|
|
15:35:37,338 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.46799999999348074. input_tokens=567, output_tokens=0
|
|
15:35:37,358 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.390999999945052. input_tokens=1308, output_tokens=0
|
|
15:35:37,404 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9529999999795109. input_tokens=566, output_tokens=0
|
|
15:35:37,420 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5940000000409782. input_tokens=456, output_tokens=0
|
|
15:35:37,448 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9529999999795109. input_tokens=414, output_tokens=0
|
|
15:35:37,470 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.125. input_tokens=465, output_tokens=0
|
|
15:35:37,512 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8279999999795109. input_tokens=769, output_tokens=0
|
|
15:35:37,536 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8429999999934807. input_tokens=480, output_tokens=0
|
|
15:35:37,560 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.25. input_tokens=243, output_tokens=0
|
|
15:35:37,581 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,584 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0309999999590218. input_tokens=539, output_tokens=0
|
|
15:35:37,599 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,622 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.327999999979511. input_tokens=448, output_tokens=0
|
|
15:35:37,637 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,643 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.140999999945052. input_tokens=749, output_tokens=0
|
|
15:35:37,666 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.077999999979511. input_tokens=575, output_tokens=0
|
|
15:35:37,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,695 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,710 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.077999999979511. input_tokens=564, output_tokens=0
|
|
15:35:37,727 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,771 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3599999999860302. input_tokens=635, output_tokens=0
|
|
15:35:37,790 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,797 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3440000000409782. input_tokens=462, output_tokens=0
|
|
15:35:37,815 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,816 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,875 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4690000000409782. input_tokens=458, output_tokens=0
|
|
15:35:37,890 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,910 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,927 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5620000000344589. input_tokens=285, output_tokens=0
|
|
15:35:37,993 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:37,998 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,7 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,13 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.46899999992456287. input_tokens=669, output_tokens=0
|
|
15:35:38,30 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,31 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,52 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.609000000054948. input_tokens=715, output_tokens=0
|
|
15:35:38,69 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,74 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9679999999934807. input_tokens=477, output_tokens=0
|
|
15:35:38,91 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4690000000409782. input_tokens=1221, output_tokens=0
|
|
15:35:38,125 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,135 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9690000000409782. input_tokens=363, output_tokens=0
|
|
15:35:38,173 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8440000000409782. input_tokens=370, output_tokens=0
|
|
15:35:38,221 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,235 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,239 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.4070000000065193. input_tokens=831, output_tokens=0
|
|
15:35:38,270 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.7969999999040738. input_tokens=358, output_tokens=0
|
|
15:35:38,309 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,333 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,339 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,374 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.5940000000409782. input_tokens=223, output_tokens=0
|
|
15:35:38,392 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,395 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0. input_tokens=575, output_tokens=0
|
|
15:35:38,412 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0929999999934807. input_tokens=585, output_tokens=0
|
|
15:35:38,459 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9679999999934807. input_tokens=254, output_tokens=0
|
|
15:35:38,474 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,480 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.875. input_tokens=649, output_tokens=0
|
|
15:35:38,496 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.062000000034459. input_tokens=459, output_tokens=0
|
|
15:35:38,513 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8439999999245629. input_tokens=385, output_tokens=0
|
|
15:35:38,538 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,567 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.4529999999795109. input_tokens=607, output_tokens=0
|
|
15:35:38,584 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,588 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0. input_tokens=542, output_tokens=0
|
|
15:35:38,628 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0940000000409782. input_tokens=1326, output_tokens=0
|
|
15:35:38,644 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9219999999040738. input_tokens=408, output_tokens=0
|
|
15:35:38,660 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8429999999934807. input_tokens=614, output_tokens=0
|
|
15:35:38,681 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,682 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,689 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0320000000065193. input_tokens=439, output_tokens=0
|
|
15:35:38,712 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.6720000000204891. input_tokens=328, output_tokens=0
|
|
15:35:38,752 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.859000000054948. input_tokens=494, output_tokens=0
|
|
15:35:38,783 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,790 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5309999999590218. input_tokens=303, output_tokens=0
|
|
15:35:38,811 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0320000000065193. input_tokens=332, output_tokens=0
|
|
15:35:38,827 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,845 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,864 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,883 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9370000000344589. input_tokens=351, output_tokens=0
|
|
15:35:38,902 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2650000000139698. input_tokens=559, output_tokens=0
|
|
15:35:38,918 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,941 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,951 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:38,966 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5. input_tokens=401, output_tokens=0
|
|
15:35:38,982 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,24 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9059999999590218. input_tokens=660, output_tokens=0
|
|
15:35:39,39 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,44 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5160000000614673. input_tokens=285, output_tokens=0
|
|
15:35:39,83 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,92 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,104 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,111 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,141 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,170 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,171 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,181 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,207 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,240 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.172000000020489. input_tokens=786, output_tokens=0
|
|
15:35:39,263 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.5939999999245629. input_tokens=428, output_tokens=0
|
|
15:35:39,300 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.875. input_tokens=581, output_tokens=0
|
|
15:35:39,320 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.172000000020489. input_tokens=453, output_tokens=0
|
|
15:35:39,358 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,361 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.6099999999860302. input_tokens=377, output_tokens=0
|
|
15:35:39,381 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.187000000034459. input_tokens=571, output_tokens=0
|
|
15:35:39,415 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,427 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0310000000754371. input_tokens=663, output_tokens=0
|
|
15:35:39,521 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0939999999245629. input_tokens=324, output_tokens=0
|
|
15:35:39,537 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:39,541 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0. input_tokens=475, output_tokens=0
|
|
15:35:39,578 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0. input_tokens=421, output_tokens=0
|
|
15:35:39,598 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8130000000819564. input_tokens=295, output_tokens=0
|
|
15:35:39,614 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.8129999999655411. input_tokens=488, output_tokens=0
|
|
15:35:39,640 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.3599999999860302. input_tokens=348, output_tokens=0
|
|
15:35:39,659 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.9839999999385327. input_tokens=309, output_tokens=0
|
|
15:35:39,675 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.077999999979511. input_tokens=310, output_tokens=0
|
|
15:35:39,695 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.015999999945052. input_tokens=300, output_tokens=0
|
|
15:35:39,730 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.2040000000270084. input_tokens=552, output_tokens=0
|
|
15:35:39,749 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 0.859000000054948. input_tokens=296, output_tokens=0
|
|
15:35:39,794 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.0940000000409782. input_tokens=527, output_tokens=0
|
|
15:35:39,973 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 1.1410000000614673. input_tokens=628, output_tokens=0
|
|
15:35:41,425 httpx INFO HTTP Request: POST https://api.openai.com/v1/embeddings "HTTP/1.1 200 OK"
|
|
15:35:42,62 graphrag.llm.base.rate_limiting_llm INFO perf - llm.embedding "Process" with 0 retries took 5.10999999998603. input_tokens=411, output_tokens=0
|
|
15:35:42,109 datashaper.workflow.workflow INFO executing verb drop
|
|
15:35:42,124 datashaper.workflow.workflow INFO executing verb filter
|
|
15:35:42,160 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_entities.parquet
|
|
15:35:42,717 graphrag.index.run.workflow INFO dependencies for create_final_nodes: ['create_base_entity_graph']
|
|
15:35:42,718 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
15:35:43,23 datashaper.workflow.workflow INFO executing verb layout_graph
|
|
15:35:59,227 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
15:35:59,786 datashaper.workflow.workflow INFO executing verb unpack_graph
|
|
15:36:00,532 datashaper.workflow.workflow INFO executing verb drop
|
|
15:36:00,551 datashaper.workflow.workflow INFO executing verb filter
|
|
15:36:00,673 datashaper.workflow.workflow INFO executing verb select
|
|
15:36:00,691 datashaper.workflow.workflow INFO executing verb snapshot
|
|
15:36:00,711 datashaper.workflow.workflow INFO executing verb rename
|
|
15:36:00,728 datashaper.workflow.workflow INFO executing verb convert
|
|
15:36:00,790 datashaper.workflow.workflow INFO executing verb join
|
|
15:36:00,825 datashaper.workflow.workflow INFO executing verb rename
|
|
15:36:00,829 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_nodes.parquet
|
|
15:36:01,545 graphrag.index.run.workflow INFO dependencies for create_final_communities: ['create_base_entity_graph']
|
|
15:36:01,545 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
15:36:01,818 datashaper.workflow.workflow INFO executing verb create_final_communities
|
|
15:36:02,937 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_communities.parquet
|
|
15:36:03,148 graphrag.index.run.workflow INFO dependencies for create_final_relationships: ['create_base_entity_graph', 'create_final_nodes']
|
|
15:36:03,149 graphrag.utils.storage INFO read table from storage: create_base_entity_graph.parquet
|
|
15:36:03,393 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
15:36:03,632 datashaper.workflow.workflow INFO executing verb create_final_relationships_pre_embedding
|
|
15:36:04,196 datashaper.workflow.workflow INFO executing verb create_final_relationships_post_embedding
|
|
15:36:04,206 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_relationships.parquet
|
|
15:36:04,440 graphrag.index.run.workflow INFO dependencies for create_final_text_units: ['create_final_relationships', 'create_final_entities', 'create_base_text_units']
|
|
15:36:04,453 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
15:36:04,460 graphrag.utils.storage INFO read table from storage: create_final_entities.parquet
|
|
15:36:04,556 graphrag.utils.storage INFO read table from storage: create_base_text_units.parquet
|
|
15:36:04,604 datashaper.workflow.workflow INFO executing verb create_final_text_units_pre_embedding
|
|
15:36:04,657 datashaper.workflow.workflow INFO executing verb select
|
|
15:36:04,659 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_text_units.parquet
|
|
15:36:04,873 graphrag.index.run.workflow INFO dependencies for create_final_community_reports: ['create_final_relationships', 'create_final_nodes']
|
|
15:36:04,874 graphrag.utils.storage INFO read table from storage: create_final_relationships.parquet
|
|
15:36:04,880 graphrag.utils.storage INFO read table from storage: create_final_nodes.parquet
|
|
15:36:05,92 datashaper.workflow.workflow INFO executing verb prepare_community_reports_nodes
|
|
15:36:05,180 datashaper.workflow.workflow INFO executing verb prepare_community_reports_edges
|
|
15:36:05,224 datashaper.workflow.workflow INFO executing verb restore_community_hierarchy
|
|
15:36:05,269 datashaper.workflow.workflow INFO executing verb prepare_community_reports
|
|
15:36:05,270 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=3 => 1612
|
|
15:36:05,305 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=2 => 1612
|
|
15:36:05,530 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=1 => 1612
|
|
15:36:05,945 graphrag.index.verbs.graph.report.prepare_community_reports INFO Number of nodes at level=0 => 1612
|
|
15:36:06,200 datashaper.workflow.workflow INFO executing verb create_community_reports
|
|
15:36:19,967 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:19,970 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 13.75. input_tokens=2062, output_tokens=500
|
|
15:36:25,730 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:25,732 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.53200000000652. input_tokens=2123, output_tokens=780
|
|
15:36:33,390 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:33,393 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.171999999904074. input_tokens=4215, output_tokens=846
|
|
15:36:37,21 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:37,23 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.796999999904074. input_tokens=4992, output_tokens=949
|
|
15:36:53,99 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:53,103 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.906000000075437. input_tokens=2216, output_tokens=544
|
|
15:36:53,245 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:53,247 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.109000000054948. input_tokens=2182, output_tokens=530
|
|
15:36:53,582 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:53,590 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.406000000075437. input_tokens=2102, output_tokens=639
|
|
15:36:53,797 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:53,799 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.719000000040978. input_tokens=2148, output_tokens=648
|
|
15:36:55,677 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:55,678 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.484000000054948. input_tokens=2387, output_tokens=564
|
|
15:36:57,60 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:57,62 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.90700000000652. input_tokens=2394, output_tokens=614
|
|
15:36:57,557 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:57,559 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.468999999924563. input_tokens=3916, output_tokens=711
|
|
15:36:57,826 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:57,828 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.67200000002049. input_tokens=2282, output_tokens=670
|
|
15:36:58,421 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:58,424 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.29700000002049. input_tokens=4900, output_tokens=659
|
|
15:36:58,770 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:58,773 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.640999999945052. input_tokens=2398, output_tokens=672
|
|
15:36:58,909 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:36:58,914 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.75. input_tokens=2443, output_tokens=779
|
|
15:37:00,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:00,323 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.15700000000652. input_tokens=2556, output_tokens=722
|
|
15:37:00,466 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:00,469 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.359000000054948. input_tokens=2570, output_tokens=850
|
|
15:37:00,848 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:00,877 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.780999999959022. input_tokens=2223, output_tokens=602
|
|
15:37:01,402 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:01,404 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.29599999997299. input_tokens=2702, output_tokens=751
|
|
15:37:02,532 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:02,538 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.436999999918044. input_tokens=4786, output_tokens=832
|
|
15:37:02,840 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:02,842 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.67200000002049. input_tokens=2944, output_tokens=814
|
|
15:37:03,9 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:03,11 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.859000000054948. input_tokens=2981, output_tokens=816
|
|
15:37:04,913 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:04,918 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.79700000002049. input_tokens=2191, output_tokens=690
|
|
15:37:05,98 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:05,100 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.984000000054948. input_tokens=4061, output_tokens=859
|
|
15:37:05,512 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:05,514 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.375. input_tokens=5097, output_tokens=908
|
|
15:37:10,572 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:10,574 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.483999999938533. input_tokens=2041, output_tokens=386
|
|
15:37:10,846 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:10,848 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 12.078000000095926. input_tokens=2152, output_tokens=409
|
|
15:37:10,901 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:10,903 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 13.34299999999348. input_tokens=2064, output_tokens=377
|
|
15:37:13,186 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:13,190 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 35.98499999998603. input_tokens=7813, output_tokens=837
|
|
15:37:14,727 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:14,728 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 13.844000000040978. input_tokens=2062, output_tokens=462
|
|
15:37:14,733 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:14,735 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.56299999996554. input_tokens=3928, output_tokens=913
|
|
15:37:16,23 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:16,24 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.18799999996554. input_tokens=2066, output_tokens=386
|
|
15:37:18,271 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:18,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.468999999924563. input_tokens=4474, output_tokens=861
|
|
15:37:19,3 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:19,10 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.93700000003446. input_tokens=3188, output_tokens=769
|
|
15:37:19,754 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:19,756 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.5. input_tokens=2293, output_tokens=617
|
|
15:37:20,112 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:20,115 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.68799999996554. input_tokens=2049, output_tokens=503
|
|
15:37:21,493 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:21,495 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.015999999945052. input_tokens=2285, output_tokens=697
|
|
15:37:21,909 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:21,917 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 11.061999999918044. input_tokens=2090, output_tokens=342
|
|
15:37:22,272 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:22,273 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.593999999924563. input_tokens=3346, output_tokens=872
|
|
15:37:23,341 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:23,344 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.438000000081956. input_tokens=5738, output_tokens=751
|
|
15:37:23,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:23,860 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.265999999945052. input_tokens=3092, output_tokens=953
|
|
15:37:26,46 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:26,48 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.531000000075437. input_tokens=2388, output_tokens=668
|
|
15:37:28,658 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:28,660 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.125. input_tokens=3959, output_tokens=757
|
|
15:37:28,941 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:28,945 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.843999999924563. input_tokens=2579, output_tokens=797
|
|
15:37:29,412 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:29,413 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.39000000001397. input_tokens=2106, output_tokens=610
|
|
15:37:31,481 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:31,483 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.56299999996554. input_tokens=3426, output_tokens=823
|
|
15:37:33,52 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:33,55 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.469000000040978. input_tokens=4848, output_tokens=681
|
|
15:37:33,176 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:33,179 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.84400000004098. input_tokens=3650, output_tokens=761
|
|
15:37:33,602 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:33,603 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.688000000081956. input_tokens=2257, output_tokens=520
|
|
15:37:34,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:34,169 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.32799999997951. input_tokens=4295, output_tokens=738
|
|
15:37:38,209 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:38,211 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.79700000002049. input_tokens=3875, output_tokens=893
|
|
15:37:41,685 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:41,687 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.5. input_tokens=3722, output_tokens=700
|
|
15:37:42,872 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:42,873 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.155999999959022. input_tokens=2132, output_tokens=839
|
|
15:37:45,335 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:45,337 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 68.25. input_tokens=2927, output_tokens=700
|
|
15:37:48,266 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:37:48,267 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 71.09399999992456. input_tokens=2232, output_tokens=723
|
|
15:38:03,498 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:03,499 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.172000000020489. input_tokens=2034, output_tokens=400
|
|
15:38:04,507 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:04,508 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.07799999997951. input_tokens=2279, output_tokens=510
|
|
15:38:05,307 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:05,309 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.98499999998603. input_tokens=2046, output_tokens=432
|
|
15:38:05,401 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:05,403 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.983999999938533. input_tokens=2255, output_tokens=537
|
|
15:38:05,424 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:05,426 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.06200000003446. input_tokens=2845, output_tokens=706
|
|
15:38:06,195 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:06,199 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.79700000002049. input_tokens=2234, output_tokens=499
|
|
15:38:09,147 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:09,149 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.73499999998603. input_tokens=2126, output_tokens=447
|
|
15:38:09,667 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:09,668 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.234000000054948. input_tokens=3474, output_tokens=683
|
|
15:38:10,75 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:10,77 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.68700000003446. input_tokens=2707, output_tokens=693
|
|
15:38:10,547 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:10,549 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.156000000075437. input_tokens=2696, output_tokens=725
|
|
15:38:10,712 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:10,724 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.359000000054948. input_tokens=2632, output_tokens=728
|
|
15:38:10,878 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:10,879 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.45299999997951. input_tokens=2764, output_tokens=732
|
|
15:38:11,63 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:11,65 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.68799999996554. input_tokens=2393, output_tokens=738
|
|
15:38:11,742 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:11,745 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.390999999945052. input_tokens=5811, output_tokens=759
|
|
15:38:12,264 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:12,266 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.921999999904074. input_tokens=2656, output_tokens=684
|
|
15:38:12,825 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:12,826 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.43700000003446. input_tokens=3370, output_tokens=771
|
|
15:38:13,380 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:13,384 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.0. input_tokens=2665, output_tokens=831
|
|
15:38:13,795 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:13,797 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.375. input_tokens=3441, output_tokens=835
|
|
15:38:15,324 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:15,326 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.89000000001397. input_tokens=2445, output_tokens=700
|
|
15:38:17,12 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:17,15 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.640999999945052. input_tokens=2337, output_tokens=650
|
|
15:38:17,533 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:17,538 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.17099999997299. input_tokens=2380, output_tokens=660
|
|
15:38:18,26 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:18,34 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.686999999918044. input_tokens=5614, output_tokens=655
|
|
15:38:18,295 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:18,298 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.92200000002049. input_tokens=3417, output_tokens=674
|
|
15:38:20,795 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:20,798 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.39100000006147. input_tokens=2678, output_tokens=768
|
|
15:38:21,654 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:21,657 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.34299999999348. input_tokens=2242, output_tokens=645
|
|
15:38:22,348 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:22,349 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.844000000040978. input_tokens=2127, output_tokens=501
|
|
15:38:22,900 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:22,905 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.57799999997951. input_tokens=2674, output_tokens=926
|
|
15:38:24,368 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:24,375 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 14.702999999979511. input_tokens=2108, output_tokens=532
|
|
15:38:25,857 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:25,859 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 14.797000000020489. input_tokens=2148, output_tokens=531
|
|
15:38:26,280 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:26,281 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.858999999938533. input_tokens=2884, output_tokens=754
|
|
15:38:27,782 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:27,784 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.061999999918044. input_tokens=2531, output_tokens=675
|
|
15:38:28,727 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:28,729 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.17200000002049. input_tokens=2351, output_tokens=617
|
|
15:38:29,210 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:29,212 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.70299999997951. input_tokens=2951, output_tokens=655
|
|
15:38:30,491 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:30,495 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.73499999998603. input_tokens=2339, output_tokens=592
|
|
15:38:32,391 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:32,395 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.06299999996554. input_tokens=2462, output_tokens=695
|
|
15:38:32,501 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:32,504 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.484000000054948. input_tokens=2534, output_tokens=627
|
|
15:38:33,459 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:33,461 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.625. input_tokens=2090, output_tokens=550
|
|
15:38:33,595 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:33,596 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.219000000040978. input_tokens=2634, output_tokens=791
|
|
15:38:34,336 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:34,338 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.125. input_tokens=2914, output_tokens=867
|
|
15:38:35,58 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:35,59 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.18799999996554. input_tokens=2758, output_tokens=667
|
|
15:38:36,481 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:36,483 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.18799999996554. input_tokens=2513, output_tokens=698
|
|
15:38:37,265 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:37,266 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.125. input_tokens=2468, output_tokens=677
|
|
15:38:37,296 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:37,298 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.891000000061467. input_tokens=2272, output_tokens=1171
|
|
15:38:39,932 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:39,938 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.67200000002049. input_tokens=2420, output_tokens=779
|
|
15:38:40,112 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:40,118 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.765999999945052. input_tokens=2250, output_tokens=746
|
|
15:38:41,165 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:41,177 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.312000000034459. input_tokens=3105, output_tokens=604
|
|
15:38:41,761 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:41,765 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.968999999924563. input_tokens=3874, output_tokens=757
|
|
15:38:42,279 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:42,281 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.375. input_tokens=2440, output_tokens=774
|
|
15:38:42,852 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:42,854 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.76600000006147. input_tokens=2224, output_tokens=847
|
|
15:38:42,887 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:42,897 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.093999999924563. input_tokens=2295, output_tokens=568
|
|
15:38:43,24 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:43,26 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.73499999998603. input_tokens=2724, output_tokens=685
|
|
15:38:43,263 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:43,268 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.73499999998603. input_tokens=2374, output_tokens=660
|
|
15:38:43,495 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:43,504 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.469000000040978. input_tokens=2653, output_tokens=793
|
|
15:38:43,660 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:43,661 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.875. input_tokens=2954, output_tokens=650
|
|
15:38:47,200 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:47,206 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.483999999938533. input_tokens=2447, output_tokens=642
|
|
15:38:48,658 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:48,659 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.155999999959022. input_tokens=2242, output_tokens=547
|
|
15:38:50,222 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:50,223 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 8.453000000095926. input_tokens=2049, output_tokens=338
|
|
15:38:51,614 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:51,617 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.10999999998603. input_tokens=6359, output_tokens=730
|
|
15:38:52,955 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:52,956 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.89000000001397. input_tokens=2390, output_tokens=557
|
|
15:38:53,382 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:53,383 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.780999999959022. input_tokens=2325, output_tokens=635
|
|
15:38:53,496 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:53,498 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.844000000040978. input_tokens=2573, output_tokens=824
|
|
15:38:54,358 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:54,361 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.98499999998603. input_tokens=2549, output_tokens=793
|
|
15:38:55,222 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:55,224 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.891000000061467. input_tokens=2435, output_tokens=622
|
|
15:38:56,59 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:56,61 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.79700000002049. input_tokens=2661, output_tokens=686
|
|
15:38:57,636 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:57,650 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.25. input_tokens=3368, output_tokens=750
|
|
15:38:58,43 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:58,45 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.031000000075437. input_tokens=2030, output_tokens=383
|
|
15:38:59,940 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:38:59,951 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.84299999999348. input_tokens=2101, output_tokens=665
|
|
15:39:00,198 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:00,199 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.70299999997951. input_tokens=2091, output_tokens=593
|
|
15:39:01,781 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:01,784 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.5. input_tokens=3205, output_tokens=716
|
|
15:39:01,988 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:01,990 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.78200000000652. input_tokens=5037, output_tokens=859
|
|
15:39:02,944 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:02,946 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.640999999945052. input_tokens=3403, output_tokens=750
|
|
15:39:03,261 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:03,263 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 26.780999999959022. input_tokens=3114, output_tokens=783
|
|
15:39:05,491 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:05,494 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.03200000000652. input_tokens=5922, output_tokens=808
|
|
15:39:05,629 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:05,633 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.45299999997951. input_tokens=3190, output_tokens=867
|
|
15:39:05,662 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:05,664 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.39000000001397. input_tokens=2099, output_tokens=502
|
|
15:39:06,451 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:06,455 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.608999999938533. input_tokens=5270, output_tokens=858
|
|
15:39:08,23 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:08,26 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.35999999998603. input_tokens=2148, output_tokens=700
|
|
15:39:08,939 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:08,941 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.718999999924563. input_tokens=2820, output_tokens=682
|
|
15:39:08,983 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:09,198 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.53200000000652. input_tokens=9537, output_tokens=915
|
|
15:39:09,282 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:09,284 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.07799999997951. input_tokens=2582, output_tokens=775
|
|
15:39:10,321 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:10,322 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 12.672000000020489. input_tokens=2128, output_tokens=389
|
|
15:39:12,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:12,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.094000000040978. input_tokens=2502, output_tokens=749
|
|
15:39:13,483 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:13,484 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 11.5. input_tokens=2064, output_tokens=366
|
|
15:39:15,410 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:15,412 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.51500000001397. input_tokens=5171, output_tokens=970
|
|
15:39:17,561 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:17,564 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.60999999998603. input_tokens=6652, output_tokens=837
|
|
15:39:18,322 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:18,866 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:18,870 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.79700000002049. input_tokens=2711, output_tokens=799
|
|
15:39:19,220 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:19,223 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.0. input_tokens=6191, output_tokens=824
|
|
15:39:19,873 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:19,874 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.92200000002049. input_tokens=2061, output_tokens=394
|
|
15:39:20,915 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.405999999959022. input_tokens=3113, output_tokens=842
|
|
15:39:21,355 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:21,358 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 41.42200000002049. input_tokens=5795, output_tokens=846
|
|
15:39:21,703 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:21,705 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.5. input_tokens=2337, output_tokens=642
|
|
15:39:23,323 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:23,325 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.96799999999348. input_tokens=2246, output_tokens=589
|
|
15:39:24,810 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:24,813 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 16.79700000002049. input_tokens=2289, output_tokens=534
|
|
15:39:25,344 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:25,346 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.73400000005495. input_tokens=2933, output_tokens=774
|
|
15:39:25,376 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:25,378 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.594000000040978. input_tokens=5873, output_tokens=863
|
|
15:39:26,90 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:26,92 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.828000000095926. input_tokens=3733, output_tokens=851
|
|
15:39:26,118 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:26,121 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.06299999996554. input_tokens=2969, output_tokens=767
|
|
15:39:26,250 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:26,252 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.31200000003446. input_tokens=2306, output_tokens=629
|
|
15:39:28,469 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:28,473 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.188000000081956. input_tokens=2473, output_tokens=716
|
|
15:39:30,810 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:30,813 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.5. input_tokens=3927, output_tokens=758
|
|
15:39:31,72 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:31,78 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 7.75. input_tokens=2033, output_tokens=377
|
|
15:39:31,468 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:31,470 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.813000000081956. input_tokens=2768, output_tokens=796
|
|
15:39:31,613 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:31,618 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.98499999998603. input_tokens=3137, output_tokens=715
|
|
15:39:32,938 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:32,940 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.45299999997951. input_tokens=2584, output_tokens=688
|
|
15:39:33,366 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:33,367 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.890999999945052. input_tokens=2582, output_tokens=660
|
|
15:39:34,150 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:34,153 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.70299999997951. input_tokens=2355, output_tokens=759
|
|
15:39:34,696 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:34,699 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.14000000001397. input_tokens=2677, output_tokens=685
|
|
15:39:34,819 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:34,824 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.625. input_tokens=5150, output_tokens=787
|
|
15:39:35,451 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:35,453 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.51500000001397. input_tokens=6147, output_tokens=902
|
|
15:39:35,786 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:35,787 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 10.436999999918044. input_tokens=2116, output_tokens=368
|
|
15:39:37,83 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:37,85 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 15.71799999999348. input_tokens=2050, output_tokens=420
|
|
15:39:37,320 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:37,323 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.82799999997951. input_tokens=4141, output_tokens=887
|
|
15:39:38,100 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:38,112 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.70400000002701. input_tokens=2612, output_tokens=720
|
|
15:39:41,8 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:41,11 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.780999999959022. input_tokens=3579, output_tokens=704
|
|
15:39:41,221 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:41,223 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.516000000061467. input_tokens=2135, output_tokens=749
|
|
15:39:44,697 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:44,700 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.82799999997951. input_tokens=2804, output_tokens=737
|
|
15:39:44,962 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:44,964 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.04700000002049. input_tokens=2927, output_tokens=814
|
|
15:39:49,337 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:49,340 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.469000000040978. input_tokens=2770, output_tokens=873
|
|
15:39:52,966 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:39:52,969 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.156000000075437. input_tokens=2348, output_tokens=776
|
|
15:40:06,209 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:06,211 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 12.937000000034459. input_tokens=2176, output_tokens=483
|
|
15:40:10,823 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:10,824 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 17.483999999938533. input_tokens=2532, output_tokens=642
|
|
15:40:13,445 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:13,447 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.17200000002049. input_tokens=2173, output_tokens=755
|
|
15:40:14,117 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:14,120 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 20.82900000002701. input_tokens=2281, output_tokens=606
|
|
15:40:14,645 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:14,647 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.296999999904074. input_tokens=2211, output_tokens=580
|
|
15:40:15,212 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:15,217 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.92200000002049. input_tokens=4822, output_tokens=837
|
|
15:40:15,771 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:15,773 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 22.515999999945052. input_tokens=7191, output_tokens=831
|
|
15:40:16,682 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:16,858 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.54700000002049. input_tokens=2093, output_tokens=520
|
|
15:40:17,140 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:17,143 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.843999999924563. input_tokens=5629, output_tokens=922
|
|
15:40:17,538 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:17,543 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.20299999997951. input_tokens=4419, output_tokens=773
|
|
15:40:17,965 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:17,966 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 24.688000000081956. input_tokens=2632, output_tokens=688
|
|
15:40:18,703 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:18,706 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 25.34299999999348. input_tokens=2176, output_tokens=635
|
|
15:40:20,545 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:20,552 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.234000000054948. input_tokens=2492, output_tokens=646
|
|
15:40:20,833 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:20,836 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 27.530999999959022. input_tokens=2708, output_tokens=652
|
|
15:40:21,415 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:21,417 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.155999999959022. input_tokens=4562, output_tokens=816
|
|
15:40:21,936 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:21,938 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 28.68799999996554. input_tokens=2467, output_tokens=691
|
|
15:40:23,166 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:23,175 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 29.93700000003446. input_tokens=2282, output_tokens=723
|
|
15:40:23,381 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:23,383 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.04700000002049. input_tokens=2148, output_tokens=740
|
|
15:40:24,155 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:24,157 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 30.82799999997951. input_tokens=6127, output_tokens=764
|
|
15:40:24,235 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:24,247 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.030999999959022. input_tokens=7943, output_tokens=760
|
|
15:40:24,701 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:24,704 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 31.42200000002049. input_tokens=3795, output_tokens=779
|
|
15:40:25,173 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:25,175 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.969000000040978. input_tokens=2358, output_tokens=607
|
|
15:40:25,684 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:25,687 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 32.45299999997951. input_tokens=5232, output_tokens=913
|
|
15:40:26,720 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:26,723 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 33.453000000095926. input_tokens=5172, output_tokens=698
|
|
15:40:27,828 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:27,830 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.5. input_tokens=3525, output_tokens=719
|
|
15:40:31,9 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:31,17 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.70299999997951. input_tokens=4631, output_tokens=797
|
|
15:40:33,153 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:33,156 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 19.030999999959022. input_tokens=3824, output_tokens=821
|
|
15:40:34,6 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:34,12 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 23.18799999996554. input_tokens=9217, output_tokens=643
|
|
15:40:35,341 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:35,343 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 18.203000000095926. input_tokens=7697, output_tokens=675
|
|
15:40:38,570 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:38,573 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 21.70299999997951. input_tokens=7876, output_tokens=829
|
|
15:40:47,449 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:47,463 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 34.01500000001397. input_tokens=3016, output_tokens=796
|
|
15:40:51,314 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:51,420 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.20299999997951. input_tokens=8715, output_tokens=772
|
|
15:40:51,466 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:51,469 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 36.828000000095926. input_tokens=7380, output_tokens=841
|
|
15:40:53,232 httpx INFO HTTP Request: POST https://api.openai.com/v1/chat/completions "HTTP/1.1 200 OK"
|
|
15:40:53,234 graphrag.llm.base.rate_limiting_llm INFO perf - llm.chat "create_community_report" with 0 retries took 37.46900000004098. input_tokens=9527, output_tokens=857
|
|
15:40:53,283 datashaper.workflow.workflow INFO executing verb window
|
|
15:40:53,286 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_community_reports.parquet
|
|
15:40:53,540 graphrag.index.run.workflow INFO dependencies for create_base_documents: ['create_final_text_units']
|
|
15:40:53,540 graphrag.utils.storage INFO read table from storage: create_final_text_units.parquet
|
|
15:40:53,593 datashaper.workflow.workflow INFO executing verb unroll
|
|
15:40:53,619 datashaper.workflow.workflow INFO executing verb select
|
|
15:40:53,667 datashaper.workflow.workflow INFO executing verb rename
|
|
15:40:53,696 datashaper.workflow.workflow INFO executing verb join
|
|
15:40:53,725 datashaper.workflow.workflow INFO executing verb aggregate_override
|
|
15:40:53,752 datashaper.workflow.workflow INFO executing verb join
|
|
15:40:53,782 datashaper.workflow.workflow INFO executing verb rename
|
|
15:40:53,807 datashaper.workflow.workflow INFO executing verb convert
|
|
15:40:53,862 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_base_documents.parquet
|
|
15:40:54,88 graphrag.index.run.workflow INFO dependencies for create_final_documents: ['create_base_documents']
|
|
15:40:54,89 graphrag.utils.storage INFO read table from storage: create_base_documents.parquet
|
|
15:40:54,144 datashaper.workflow.workflow INFO executing verb rename
|
|
15:40:54,145 graphrag.index.emit.parquet_table_emitter INFO emitting parquet table create_final_documents.parquet
|
|
15:40:54,257 graphrag.index.cli INFO All workflows completed successfully.
|
|
|