跳至主要内容

SearchApi

此笔记本展示了如何使用 SearchApi 搜索网络的示例。请访问 https://www.searchapi.io/ 注册免费帐户并获取 API 密钥。

import os

os.environ["SEARCHAPI_API_KEY"] = ""
from langchain_community.utilities import SearchApiAPIWrapper
API 参考:SearchApiAPIWrapper
search = SearchApiAPIWrapper()
search.run("Obama's first name?")
'Barack Hussein Obama II'

作为自问自答搜索链的一部分使用

os.environ["OPENAI_API_KEY"] = ""
from langchain.agents import AgentType, initialize_agent
from langchain_community.utilities import SearchApiAPIWrapper
from langchain_core.tools import Tool
from langchain_openai import OpenAI

llm = OpenAI(temperature=0)
search = SearchApiAPIWrapper()
tools = [
Tool(
name="Intermediate Answer",
func=search.run,
description="useful for when you need to ask with search",
)
]

self_ask_with_search = initialize_agent(
tools, llm, agent=AgentType.SELF_ASK_WITH_SEARCH, verbose=True
)
self_ask_with_search.run("Who lived longer: Plato, Socrates, or Aristotle?")


> Entering new AgentExecutor chain...
 Yes.
Follow up: How old was Plato when he died?
Intermediate answer: eighty
Follow up: How old was Socrates when he died?
Intermediate answer: | Socrates |
| -------- |
| Born | c. 470 BC Deme Alopece, Athens |
| Died | 399 BC (aged approximately 71) Athens |
| Cause of death | Execution by forced suicide by poisoning |
| Spouse(s) | Xanthippe, Myrto |

Follow up: How old was Aristotle when he died?
Intermediate answer: 62 years
So the final answer is: Plato

> Finished chain.
'Plato'

自定义参数

SearchApi 包装器可以自定义为使用不同的引擎,例如 Google 新闻Google 招聘Google 学术搜索 或其他可在 SearchApi 文档中找到的引擎。在执行查询时,可以传递 SearchApi 支持的所有参数。

search = SearchApiAPIWrapper(engine="google_jobs")
search.run("AI Engineer", location="Portugal", gl="pt")[0:500]
'Azure AI Engineer Be an XpanderCandidatar-meCandidatar-meCandidatar-me\n\nShare:\n\nAzure AI Engineer\n\nA área Digital Xperience da Xpand IT é uma equipa tecnológica de rápido crescimento que se concentra em tecnologias Microsoft e Mobile. A sua principal missão é fornecer soluções de software de alta qualidade que atendam às necessidades do utilizador final, num mundo tecnológico continuamente exigente e em ritmo acelerado, proporcionando a melhor experiência em termos de personalização, performance'

获取带元数据的結果

import pprint
search = SearchApiAPIWrapper(engine="google_scholar")
results = search.results("Large Language Models")
pprint.pp(results)
{'search_metadata': {'id': 'search_qVdXG2jzvrlqTzayeYoaOb8A',
'status': 'Success',
'created_at': '2023-09-25T15:22:30Z',
'request_time_taken': 3.21,
'parsing_time_taken': 0.03,
'total_time_taken': 3.24,
'request_url': 'https://scholar.google.com/scholar?q=Large+Language+Models&hl=en',
'html_url': 'https://www.searchapi.io/api/v1/searches/search_qVdXG2jzvrlqTzayeYoaOb8A.html',
'json_url': 'https://www.searchapi.io/api/v1/searches/search_qVdXG2jzvrlqTzayeYoaOb8A'},
'search_parameters': {'engine': 'google_scholar',
'q': 'Large Language Models',
'hl': 'en'},
'search_information': {'query_displayed': 'Large Language Models',
'total_results': 6420000,
'page': 1,
'time_taken_displayed': 0.06},
'organic_results': [{'position': 1,
'title': 'ChatGPT for good? On opportunities and '
'challenges of large language models for '
'education',
'data_cid': 'uthwmf2nU3EJ',
'link': 'https://www.sciencedirect.com/science/article/pii/S1041608023000195',
'publication': 'E Kasneci, K Seßler, S Küchemann, M '
'Bannert… - Learning and individual …, '
'2023 - Elsevier',
'snippet': '… state of large language models and their '
'applications. We then highlight how these '
'models can be … With regard to challenges, '
'we argue that large language models in '
'education require …',
'inline_links': {'cited_by': {'cites_id': '8166055256995715258',
'total': 410,
'link': 'https://scholar.google.com/scholar?cites=8166055256995715258&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '8166055256995715258',
'total': 10,
'link': 'https://scholar.google.com/scholar?cluster=8166055256995715258&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:uthwmf2nU3EJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'edarxiv.org',
'format': 'PDF',
'link': 'https://edarxiv.org/5er8f/download?format=pdf'},
'authors': [{'name': 'E Kasneci',
'id': 'bZVkVvoAAAAJ',
'link': 'https://scholar.google.com/citations?user=bZVkVvoAAAAJ&hl=en&oi=sra'},
{'name': 'K Seßler',
'id': 'MbMBoN4AAAAJ',
'link': 'https://scholar.google.com/citations?user=MbMBoN4AAAAJ&hl=en&oi=sra'},
{'name': 'S Küchemann',
'id': 'g1jX5QUAAAAJ',
'link': 'https://scholar.google.com/citations?user=g1jX5QUAAAAJ&hl=en&oi=sra'},
{'name': 'M Bannert',
'id': 'TjfQ8QkAAAAJ',
'link': 'https://scholar.google.com/citations?user=TjfQ8QkAAAAJ&hl=en&oi=sra'}]},
{'position': 2,
'title': 'Large language models in medicine',
'data_cid': 'Ph9AwHTmhzAJ',
'link': 'https://www.nature.com/articles/s41591-023-02448-8',
'publication': 'AJ Thirunavukarasu, DSJ Ting, K '
'Elangovan… - Nature medicine, 2023 - '
'nature.com',
'snippet': '… HuggingChat offers a free-to-access '
'chatbot with a similar interface to ChatGPT '
'but uses Large Language Model Meta AI '
'(LLaMA) as its backend model 30 . Finally, '
'cheap imitations of …',
'inline_links': {'cited_by': {'cites_id': '3497017024792502078',
'total': 25,
'link': 'https://scholar.google.com/scholar?cites=3497017024792502078&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '3497017024792502078',
'total': 3,
'link': 'https://scholar.google.com/scholar?cluster=3497017024792502078&hl=en&as_sdt=0,33'}},
'authors': [{'name': 'AJ Thirunavukarasu',
'id': '3qb1AYwAAAAJ',
'link': 'https://scholar.google.com/citations?user=3qb1AYwAAAAJ&hl=en&oi=sra'},
{'name': 'DSJ Ting',
'id': 'KbrpC8cAAAAJ',
'link': 'https://scholar.google.com/citations?user=KbrpC8cAAAAJ&hl=en&oi=sra'},
{'name': 'K Elangovan',
'id': 'BE_lVTQAAAAJ',
'link': 'https://scholar.google.com/citations?user=BE_lVTQAAAAJ&hl=en&oi=sra'}]},
{'position': 3,
'title': 'Extracting training data from large language '
'models',
'data_cid': 'mEYsWK6bWKoJ',
'link': 'https://www.usenix.org/conference/usenixsecurity21/presentation/carlini-extracting',
'publication': 'N Carlini, F Tramer, E Wallace, M '
'Jagielski… - 30th USENIX Security …, '
'2021 - usenix.org',
'snippet': '… language model trained on scrapes of the '
'public Internet, and are able to extract '
'hundreds of verbatim text sequences from the '
'model’… models are more vulnerable than '
'smaller models. …',
'inline_links': {'cited_by': {'cites_id': '12274731957504198296',
'total': 742,
'link': 'https://scholar.google.com/scholar?cites=12274731957504198296&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '12274731957504198296',
'total': 8,
'link': 'https://scholar.google.com/scholar?cluster=12274731957504198296&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:mEYsWK6bWKoJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:mEYsWK6bWKoJ:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'usenix.org',
'format': 'PDF',
'link': 'https://www.usenix.org/system/files/sec21-carlini-extracting.pdf'},
'authors': [{'name': 'N Carlini',
'id': 'q4qDvAoAAAAJ',
'link': 'https://scholar.google.com/citations?user=q4qDvAoAAAAJ&hl=en&oi=sra'},
{'name': 'F Tramer',
'id': 'ijH0-a8AAAAJ',
'link': 'https://scholar.google.com/citations?user=ijH0-a8AAAAJ&hl=en&oi=sra'},
{'name': 'E Wallace',
'id': 'SgST3LkAAAAJ',
'link': 'https://scholar.google.com/citations?user=SgST3LkAAAAJ&hl=en&oi=sra'},
{'name': 'M Jagielski',
'id': '_8rw_GMAAAAJ',
'link': 'https://scholar.google.com/citations?user=_8rw_GMAAAAJ&hl=en&oi=sra'}]},
{'position': 4,
'title': 'Emergent abilities of large language models',
'data_cid': 'hG0iVOrOguoJ',
'link': 'https://arxiv.org/abs/2206.07682',
'publication': 'J Wei, Y Tay, R Bommasani, C Raffel, B '
'Zoph… - arXiv preprint arXiv …, 2022 - '
'arxiv.org',
'snippet': 'Scaling up language models has been shown to '
'predictably improve performance and sample '
'efficiency on a wide range of downstream '
'tasks. This paper instead discusses an …',
'inline_links': {'cited_by': {'cites_id': '16898296257676733828',
'total': 621,
'link': 'https://scholar.google.com/scholar?cites=16898296257676733828&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '16898296257676733828',
'total': 12,
'link': 'https://scholar.google.com/scholar?cluster=16898296257676733828&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:hG0iVOrOguoJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:hG0iVOrOguoJ:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'arxiv.org',
'format': 'PDF',
'link': 'https://arxiv.org/pdf/2206.07682.pdf?trk=cndc-detail'},
'authors': [{'name': 'J Wei',
'id': 'wA5TK_0AAAAJ',
'link': 'https://scholar.google.com/citations?user=wA5TK_0AAAAJ&hl=en&oi=sra'},
{'name': 'Y Tay',
'id': 'VBclY_cAAAAJ',
'link': 'https://scholar.google.com/citations?user=VBclY_cAAAAJ&hl=en&oi=sra'},
{'name': 'R Bommasani',
'id': 'WMBXw1EAAAAJ',
'link': 'https://scholar.google.com/citations?user=WMBXw1EAAAAJ&hl=en&oi=sra'},
{'name': 'C Raffel',
'id': 'I66ZBYwAAAAJ',
'link': 'https://scholar.google.com/citations?user=I66ZBYwAAAAJ&hl=en&oi=sra'},
{'name': 'B Zoph',
'id': 'NL_7iTwAAAAJ',
'link': 'https://scholar.google.com/citations?user=NL_7iTwAAAAJ&hl=en&oi=sra'}]},
{'position': 5,
'title': 'A survey on evaluation of large language '
'models',
'data_cid': 'ZYohnzOz-XgJ',
'link': 'https://arxiv.org/abs/2307.03109',
'publication': 'Y Chang, X Wang, J Wang, Y Wu, K Zhu… - '
'arXiv preprint arXiv …, 2023 - arxiv.org',
'snippet': '… 3.1 Natural Language Processing Tasks … '
'the development of language models, '
'particularly large language models, was to '
'enhance performance on natural language '
'processing tasks, …',
'inline_links': {'cited_by': {'cites_id': '8717195588046785125',
'total': 31,
'link': 'https://scholar.google.com/scholar?cites=8717195588046785125&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '8717195588046785125',
'total': 3,
'link': 'https://scholar.google.com/scholar?cluster=8717195588046785125&hl=en&as_sdt=0,33'},
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:ZYohnzOz-XgJ:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'arxiv.org',
'format': 'PDF',
'link': 'https://arxiv.org/pdf/2307.03109'},
'authors': [{'name': 'X Wang',
'id': 'Q7Ieos8AAAAJ',
'link': 'https://scholar.google.com/citations?user=Q7Ieos8AAAAJ&hl=en&oi=sra'},
{'name': 'J Wang',
'id': 'YomxTXQAAAAJ',
'link': 'https://scholar.google.com/citations?user=YomxTXQAAAAJ&hl=en&oi=sra'},
{'name': 'Y Wu',
'id': 'KVeRu2QAAAAJ',
'link': 'https://scholar.google.com/citations?user=KVeRu2QAAAAJ&hl=en&oi=sra'},
{'name': 'K Zhu',
'id': 'g75dFLYAAAAJ',
'link': 'https://scholar.google.com/citations?user=g75dFLYAAAAJ&hl=en&oi=sra'}]},
{'position': 6,
'title': 'Evaluating large language models trained on '
'code',
'data_cid': '3tNvW3l5nU4J',
'link': 'https://arxiv.org/abs/2107.03374',
'publication': 'M Chen, J Tworek, H Jun, Q Yuan, HPO '
'Pinto… - arXiv preprint arXiv …, 2021 - '
'arxiv.org',
'snippet': '… We introduce Codex, a GPT language model '
'finetuned on publicly available code from '
'GitHub, and study its Python code-writing '
'capabilities. A distinct production version '
'of Codex …',
'inline_links': {'cited_by': {'cites_id': '5664817468434011102',
'total': 941,
'link': 'https://scholar.google.com/scholar?cites=5664817468434011102&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '5664817468434011102',
'total': 2,
'link': 'https://scholar.google.com/scholar?cluster=5664817468434011102&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:3tNvW3l5nU4J:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:3tNvW3l5nU4J:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'arxiv.org',
'format': 'PDF',
'link': 'https://arxiv.org/pdf/2107.03374.pdf?trk=public_post_comment-text'},
'authors': [{'name': 'M Chen',
'id': '5fU-QMwAAAAJ',
'link': 'https://scholar.google.com/citations?user=5fU-QMwAAAAJ&hl=en&oi=sra'},
{'name': 'J Tworek',
'id': 'ZPuESCQAAAAJ',
'link': 'https://scholar.google.com/citations?user=ZPuESCQAAAAJ&hl=en&oi=sra'},
{'name': 'Q Yuan',
'id': 'B059m2EAAAAJ',
'link': 'https://scholar.google.com/citations?user=B059m2EAAAAJ&hl=en&oi=sra'}]},
{'position': 7,
'title': 'Large language models in machine translation',
'data_cid': 'sY5m_Y3-0Y4J',
'link': 'http://research.google/pubs/pub33278.pdf',
'publication': 'T Brants, AC Popat, P Xu, FJ Och, J Dean '
'- 2007 - research.google',
'snippet': '… the benefits of largescale statistical '
'language modeling in ma… trillion tokens, '
'resulting in language models having up to '
'300 … is inexpensive to train on large data '
'sets and approaches the …',
'type': 'PDF',
'inline_links': {'cited_by': {'cites_id': '10291286509313494705',
'total': 737,
'link': 'https://scholar.google.com/scholar?cites=10291286509313494705&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '10291286509313494705',
'total': 31,
'link': 'https://scholar.google.com/scholar?cluster=10291286509313494705&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:sY5m_Y3-0Y4J:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:sY5m_Y3-0Y4J:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'research.google',
'format': 'PDF',
'link': 'http://research.google/pubs/pub33278.pdf'},
'authors': [{'name': 'FJ Och',
'id': 'ITGdg6oAAAAJ',
'link': 'https://scholar.google.com/citations?user=ITGdg6oAAAAJ&hl=en&oi=sra'},
{'name': 'J Dean',
'id': 'NMS69lQAAAAJ',
'link': 'https://scholar.google.com/citations?user=NMS69lQAAAAJ&hl=en&oi=sra'}]},
{'position': 8,
'title': 'A watermark for large language models',
'data_cid': 'BlSyLHT4iiEJ',
'link': 'https://arxiv.org/abs/2301.10226',
'publication': 'J Kirchenbauer, J Geiping, Y Wen, J '
'Katz… - arXiv preprint arXiv …, 2023 - '
'arxiv.org',
'snippet': '… To derive this watermark, we examine what '
'happens in the language model just before it '
'produces a probability vector. The last '
'layer of the language model outputs a vector '
'of logits l(t). …',
'inline_links': {'cited_by': {'cites_id': '2417017327887471622',
'total': 104,
'link': 'https://scholar.google.com/scholar?cites=2417017327887471622&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '2417017327887471622',
'total': 4,
'link': 'https://scholar.google.com/scholar?cluster=2417017327887471622&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:BlSyLHT4iiEJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:BlSyLHT4iiEJ:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'arxiv.org',
'format': 'PDF',
'link': 'https://arxiv.org/pdf/2301.10226.pdf?curius=1419'},
'authors': [{'name': 'J Kirchenbauer',
'id': '48GJrbsAAAAJ',
'link': 'https://scholar.google.com/citations?user=48GJrbsAAAAJ&hl=en&oi=sra'},
{'name': 'J Geiping',
'id': '206vNCEAAAAJ',
'link': 'https://scholar.google.com/citations?user=206vNCEAAAAJ&hl=en&oi=sra'},
{'name': 'Y Wen',
'id': 'oUYfjg0AAAAJ',
'link': 'https://scholar.google.com/citations?user=oUYfjg0AAAAJ&hl=en&oi=sra'},
{'name': 'J Katz',
'id': 'yPw4WjoAAAAJ',
'link': 'https://scholar.google.com/citations?user=yPw4WjoAAAAJ&hl=en&oi=sra'}]},
{'position': 9,
'title': 'ChatGPT and other large language models are '
'double-edged swords',
'data_cid': 'So0q8TRvxhYJ',
'link': 'https://pubs.rsna.org/doi/full/10.1148/radiol.230163',
'publication': 'Y Shen, L Heacock, J Elias, KD Hentel, B '
'Reig, G Shih… - Radiology, 2023 - '
'pubs.rsna.org',
'snippet': '… Large Language Models (LLMs) are deep '
'learning models trained to understand and '
'generate natural language. Recent studies '
'demonstrated that LLMs achieve great success '
'in a …',
'inline_links': {'cited_by': {'cites_id': '1641121387398204746',
'total': 231,
'link': 'https://scholar.google.com/scholar?cites=1641121387398204746&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '1641121387398204746',
'total': 3,
'link': 'https://scholar.google.com/scholar?cluster=1641121387398204746&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:So0q8TRvxhYJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33'},
'authors': [{'name': 'Y Shen',
'id': 'XaeN2zgAAAAJ',
'link': 'https://scholar.google.com/citations?user=XaeN2zgAAAAJ&hl=en&oi=sra'},
{'name': 'L Heacock',
'id': 'tYYM5IkAAAAJ',
'link': 'https://scholar.google.com/citations?user=tYYM5IkAAAAJ&hl=en&oi=sra'}]},
{'position': 10,
'title': 'Pythia: A suite for analyzing large language '
'models across training and scaling',
'data_cid': 'aaIDvsMAD8QJ',
'link': 'https://proceedings.mlr.press/v202/biderman23a.html',
'publication': 'S Biderman, H Schoelkopf… - '
'International …, 2023 - '
'proceedings.mlr.press',
'snippet': '… large language models, we prioritize '
'consistency in model … out the most '
'performance from each model. For example, we '
'… models, as it is becoming widely used for '
'the largest models, …',
'inline_links': {'cited_by': {'cites_id': '14127511396791067241',
'total': 89,
'link': 'https://scholar.google.com/scholar?cites=14127511396791067241&as_sdt=5,33&sciodt=0,33&hl=en'},
'versions': {'cluster_id': '14127511396791067241',
'total': 3,
'link': 'https://scholar.google.com/scholar?cluster=14127511396791067241&hl=en&as_sdt=0,33'},
'related_articles_link': 'https://scholar.google.com/scholar?q=related:aaIDvsMAD8QJ:scholar.google.com/&scioq=Large+Language+Models&hl=en&as_sdt=0,33',
'cached_page_link': 'https://scholar.googleusercontent.com/scholar?q=cache:aaIDvsMAD8QJ:scholar.google.com/+Large+Language+Models&hl=en&as_sdt=0,33'},
'resource': {'name': 'mlr.press',
'format': 'PDF',
'link': 'https://proceedings.mlr.press/v202/biderman23a/biderman23a.pdf'},
'authors': [{'name': 'S Biderman',
'id': 'bO7H0DAAAAAJ',
'link': 'https://scholar.google.com/citations?user=bO7H0DAAAAAJ&hl=en&oi=sra'},
{'name': 'H Schoelkopf',
'id': 'XLahYIYAAAAJ',
'link': 'https://scholar.google.com/citations?user=XLahYIYAAAAJ&hl=en&oi=sra'}]}],
'related_searches': [{'query': 'large language models machine',
'highlighted': ['machine'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=1&q=large+language+models+machine&qst=ib'},
{'query': 'large language models pruning',
'highlighted': ['pruning'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=2&q=large+language+models+pruning&qst=ib'},
{'query': 'large language models multitask learners',
'highlighted': ['multitask learners'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=3&q=large+language+models+multitask+learners&qst=ib'},
{'query': 'large language models speech recognition',
'highlighted': ['speech recognition'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=4&q=large+language+models+speech+recognition&qst=ib'},
{'query': 'large language models machine translation',
'highlighted': ['machine translation'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=5&q=large+language+models+machine+translation&qst=ib'},
{'query': 'emergent abilities of large language models',
'highlighted': ['emergent abilities of'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=6&q=emergent+abilities+of+large+language+models&qst=ir'},
{'query': 'language models privacy risks',
'highlighted': ['privacy risks'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=7&q=language+models+privacy+risks&qst=ir'},
{'query': 'language model fine tuning',
'highlighted': ['fine tuning'],
'link': 'https://scholar.google.com/scholar?hl=en&as_sdt=0,33&qsp=8&q=language+model+fine+tuning&qst=ir'}],
'pagination': {'current': 1,
'next': 'https://scholar.google.com/scholar?start=10&q=Large+Language+Models&hl=en&as_sdt=0,33',
'other_pages': {'2': 'https://scholar.google.com/scholar?start=10&q=Large+Language+Models&hl=en&as_sdt=0,33',
'3': 'https://scholar.google.com/scholar?start=20&q=Large+Language+Models&hl=en&as_sdt=0,33',
'4': 'https://scholar.google.com/scholar?start=30&q=Large+Language+Models&hl=en&as_sdt=0,33',
'5': 'https://scholar.google.com/scholar?start=40&q=Large+Language+Models&hl=en&as_sdt=0,33',
'6': 'https://scholar.google.com/scholar?start=50&q=Large+Language+Models&hl=en&as_sdt=0,33',
'7': 'https://scholar.google.com/scholar?start=60&q=Large+Language+Models&hl=en&as_sdt=0,33',
'8': 'https://scholar.google.com/scholar?start=70&q=Large+Language+Models&hl=en&as_sdt=0,33',
'9': 'https://scholar.google.com/scholar?start=80&q=Large+Language+Models&hl=en&as_sdt=0,33',
'10': 'https://scholar.google.com/scholar?start=90&q=Large+Language+Models&hl=en&as_sdt=0,33'}}}

此页面有幫助嗎?


您也可以在 GitHub 上留下详细的反馈 GitHub.