<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Open AI Archives - Good Shepherd News - Fastest Growing Religious, Free Speech &amp; Political Content</title>
	<atom:link href="https://goodshepherdmedia.net/tag/open-ai/feed/" rel="self" type="application/rss+xml" />
	<link>https://goodshepherdmedia.net/tag/open-ai/</link>
	<description>Christian, Political, ‎‏‏‎Social &#38; Legal Free Speech News &#124; Ⓒ2024 Good News Media LLC &#124; Shepherd for the Herd! God 1st Programming</description>
	<lastBuildDate>Fri, 03 Feb 2023 08:12:15 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.1</generator>

<image>
	<url>https://goodshepherdmedia.net/wp-content/uploads/2023/08/Good-Shepherd-News-Logo-150x150.png</url>
	<title>Open AI Archives - Good Shepherd News - Fastest Growing Religious, Free Speech &amp; Political Content</title>
	<link>https://goodshepherdmedia.net/tag/open-ai/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Big Tech was moving cautiously on AI. Then came ChatGPT.</title>
		<link>https://goodshepherdmedia.net/big-tech-was-moving-cautiously-on-ai-then-came-chatgpt/</link>
		
		<dc:creator><![CDATA[The Truth News]]></dc:creator>
		<pubDate>Sat, 28 Jan 2023 00:27:10 +0000</pubDate>
				<category><![CDATA[How To]]></category>
		<category><![CDATA[Tech]]></category>
		<category><![CDATA[Top Stories]]></category>
		<category><![CDATA[Zee Truthful News]]></category>
		<category><![CDATA[💻Tech History]]></category>
		<category><![CDATA[🤖 AI Artificial Intelligence]]></category>
		<category><![CDATA[🤖Open AI]]></category>
		<category><![CDATA[🤖🗣️ChatGPT]]></category>
		<category><![CDATA[Amazon Warns of ChatGPT]]></category>
		<category><![CDATA[Beware of ChatGPT]]></category>
		<category><![CDATA[Big Tech]]></category>
		<category><![CDATA[Chat GPT]]></category>
		<category><![CDATA[ChatGPT]]></category>
		<category><![CDATA[Open AI]]></category>
		<guid isPermaLink="false">https://goodshepherdmedia.net/?p=10105</guid>

					<description><![CDATA[Big Tech was moving cautiously on AI. Then came ChatGPT. Google, Facebook and Microsoft helped build the scaffolding of AI. Smaller companies are taking it to the masses, forcing Big Tech to react. What is ChatGPT? Three months before ChatGPT debuted in November, Facebook’s parent company Meta released a similar chatbot. But unlike the phenomenon [&#8230;]]]></description>
										<content:encoded><![CDATA[<h1 style="text-align: center;"><span style="font-size: 24pt;">Big Tech was moving cautiously on AI.<br />
Then came ChatGPT.</span></h1>
<div>
<blockquote>
<h3 class="font--subhead font-light offblack mb-sm pb-xxs-ns subheadline" style="text-align: center;" data-qa="subheadline"><em>Google, Facebook and Microsoft helped build the scaffolding of AI. Smaller companies are taking it to the masses, forcing Big Tech to react.</em></h3>
</blockquote>
</div>
<div>
<h1 style="text-align: center;">What is ChatGPT?</h1>
</div>
<p><iframe title="AI Unveiled: ChatGPT and Deepfake - Hear this Critical Warning!" width="640" height="360" src="https://www.youtube.com/embed/1EDqNyefUmM?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></p>
<div>Three months before ChatGPT debuted in November, Facebook’s parent company Meta released a similar chatbot. But unlike the phenomenon that ChatGPT instantly became, with more than a million users in its first five days, Meta’s Blenderbot was boring, said Meta’s chief artificial intelligence scientist, Yann LeCun.</div>
<div></div>
<div>“The reason it was boring was because it was made safe,” LeCun said last week at a forum hosted by AI consulting company Collective[i]. He blamed the tepid public response on Meta being “overly careful about content moderation,” like directing the chatbot to change the subject if a user asked about religion. ChatGPT, on the other hand, will converse about the concept of falsehoods in the Quran, write a prayer for a rabbi to deliver to Congress and compare God to a flyswatter.</div>
<div></div>
<div>ChatGPT is quickly going mainstream now that Microsoft — which recently invested billions of dollars in the company behind the chatbot, OpenAI — is working to incorporate it into its popular office software and selling access to the tool to other businesses. The surge of attention around ChatGPT is prompting pressure inside tech giants including Meta and Google to move faster, potentially sweeping safety concerns aside, according to interviews with six current and former Google and Meta employees, some of whom spoke on the condition of anonymity because they were not authorized to speak.</div>
<div></div>
<div>At Meta, employees have recently shared internal memos urging the company to speed up its AI approval process to take advantage of the latest technology, according to one of them. Google, which helped pioneer some of the technology underpinning ChatGPT, recently issued a “code red” around launching AI products and proposed a “green lane” to shorten the process of assessing and mitigating potential harms, according to a report in the New York Times.</div>
<div></div>
<div>ChatGPT, along with text-to-image tools such as DALL-E 2 and Stable Diffusion, is part of a new wave of software called generative AI. They create works of their own by drawing on patterns they’ve identified in vast troves of existing, human-created content. This technology was pioneered at big tech companies like Google that in recent years have grown more secretive, announcing new models or offering demos but keeping the full product under lock and key. Meanwhile, research labs like OpenAI rapidly launched their latest versions, raising questions about how corporate offerings, like Google’s language model LaMDA, stack up.</div>
<div></div>
<div>Tech giants have been skittish since public debacles like Microsoft’s Tay, which it took down in less than a day in 2016 after trolls prompted the bot to call for a race war, suggest Hitler was right and tweet “Jews did 9/11.” Meta defended Blenderbot and left it up after it made racist comments in August, but pulled down another AI tool, called Galactica, in November after just three days amid criticism over its inaccurate and sometimes biased summaries of scientific research.</div>
<div></div>
<div>“People feel like OpenAI is newer, fresher, more exciting and has fewer sins to pay for than these incumbent companies, and they can get away with this for now,” said a Google employee who works in AI, referring to the public’s willingness to accept ChatGPT with less scrutiny. Some top talent has jumped ship to nimbler start-ups, like OpenAI and Stable Diffusion.</div>
<div></div>
<div>Some AI ethicists fear that Big Tech’s rush to market could expose billions of people to potential harms — such as sharing inaccurate information, generating fake photos or giving students the ability to cheat on school tests — before trust and safety experts have been able to study the risks. Others in the field share OpenAI’s philosophy that releasing the tools to the public, often nominally in a “beta” phase after mitigating some predictable risks, is the only way to assess real world harms.</div>
<div></div>
<div>“The pace of progress in AI is incredibly fast, and we are always keeping an eye on making sure we have efficient review processes, but the priority is to make the right decisions, and release AI models and products that best serve our community,” said Joelle Pineau, managing director of Fundamental AI Research at Meta.</div>
<div></div>
<div>“We believe that AI is foundational and transformative technology that is incredibly useful for individuals, businesses and communities,” said Lily Lin, a Google spokesperson. “We need to consider the broader societal impacts these innovations can have. We continue to test our AI technology internally to make sure it’s helpful and safe.”</div>
<div></div>
<div>Microsoft’s chief of communications, Frank Shaw, said his company works with OpenAI to build in extra safety mitigations when it uses AI tools like DALLE-2 in its products. “Microsoft has been working for years to both advance the field of AI and publicly guide how these technologies are created and used on our platforms in responsible and ethical ways,” Shaw said.</div>
<div>OpenAI declined to comment.</div>
<div></div>
<div>The technology underlying ChatGPT isn’t necessarily better than what Google and Meta have developed, said Mark Riedl, professor of computing at Georgia Tech and an expert on machine learning. <strong>But OpenAI’s practice of releasing its language models for public use has given it a real advantage.</strong></div>
<div></div>
<div>“For the last two years they’ve been using a crowd of humans to provide feedback to GPT,” said Riedl, such as giving a “thumbs down” for an inappropriate or unsatisfactory answer, a process called “reinforcement learning from human feedback.”</div>
<div>Silicon Valley’s sudden willingness to consider taking more reputational risk arrives as tech stocks are tumbling. When Google laid off 12,000 employees last week, CEO Sundar Pichai wrote that the company had undertaken a rigorous review to focus on its highest priorities, twice referencing its early investments in AI.</div>
<div></div>
<div>A decade ago, Google was the undisputed leader in the field. It acquired the cutting edge AI lab DeepMind in 2014 and open-sourced its machine learning software TensorFlow in 2015. By 2016, Pichai pledged to transform Google into an “AI first” company.</div>
<div>The next year, Google released transformers — a pivotal piece of software architecture that made the current wave of generative AI possible.</div>
<div></div>
<div>The company kept rolling out state-of-the-art technology that propelled the entire field forward, deploying some AI breakthroughs in understanding language to improve Google search. Inside big tech companies, the system of checks and balances for vetting the ethical implications of cutting-edge AI isn’t as established as privacy or data security. Typically teams of AI researchers and engineers publish papers on their findings, incorporate their technology into the company’s existing infrastructure or develop new products, a process that can sometimes clash with other teams working on responsible AI over pressure to see innovation reach the public sooner.</div>
<div></div>
<div>Google released its AI principles in 2018, after facing employee protest over Project Maven, a contract to provide computer vision for Pentagon drones, and consumer backlash over a demo for Duplex, an AI system that would call restaurants and make a reservation without disclosing it was a bot. In August last year, Google began giving consumers access to a limited version of LaMDA through its app AI Test Kitchen. It has not yet released it fully to the general public, despite Google’s plans to do so at the end of 2022, according to former Google software engineer Blake Lemoine, who told The Washington Post that he had come to believe LaMDA was sentient.</div>
<div>[The Google engineer who thinks the company’s AI has come to life]</div>
<div></div>
<p>But the top AI talent behind these developments grew restless.</p>
<div></div>
<div>In the past year or so, top AI researchers from Google have left to launch start-ups around large language models, including Character.AI, Cohere, Adept, Inflection.AI and Inworld AI, in addition to search start-ups using similar models to develop a chat interface, such as Neeva, run by former Google executive Sridhar Ramaswamy.</div>
<div></div>
<div>Character.AI founder Noam Shazeer, who helped invent the transformer and other core machine learning architecture, said the flywheel effect of user data has been invaluable. The first time he applied user feedback to Character.AI, which allows anyone to generate chatbots based on short descriptions of real people or imaginary figures, engagement rose by more than 30 percent.</div>
<div></div>
<div>Bigger companies like Google and Microsoft are generally focused on using AI to improve their massive existing business models, said Nick Frosst, who worked at Google Brain for three years before co-founding Cohere, a Toronto-based start-up building large language models that can be customized to help businesses. One of his co-founders, Aidan Gomez, also helped invent transformers when he worked at Google.</div>
<div></div>
<div>“The space moves so quickly, it’s not surprising to me that the people leading are smaller companies,” said Frosst.</div>
<div></div>
<div>AI has been through several hype cycles over the past decade, but the furor over DALL-E and ChatGPT has reached new heights.</div>
<div></div>
<div>Soon after OpenAI released ChatGPT, tech influencers on Twitter began to predict that generative AI would spell the demise of Google search. ChatGPT delivered simple answers in an accessible way and didn’t ask users to rifle through blue links. Besides, after a quarter of a century, Google’s search interface had grown bloated with ads and marketers trying to game the system.</div>
<div></div>
<div>“Thanks to their monopoly position, the folks over at Mountain View have [let] their once-incredible search experience degenerate into a spam-ridden, SEO-fueled hellscape,” technologist Can Duruk wrote in his newsletter Margins, referring to Google’s hometown.</div>
<div>On the anonymous app Blind, tech workers posted dozens of questions about whether the Silicon Valley giant could compete.</div>
<div></div>
<div>“If Google doesn’t get their act together and start shipping, they will go down in history as the company who nurtured and trained an entire generation of machine learning researchers and engineers who went on to deploy the technology at other companies,” tweeted David Ha, a renowned research scientist who recently left Google Brain for the open source text-to-image start-up Stable Diffusion.</div>
<div></div>
<div>AI engineers still inside Google shared his frustration, employees say. For years, employees had sent memos about incorporating chat functions into search, viewing it as an obvious evolution, according to employees. But they also understood that Google had justifiable reasons not to be hasty about switching up its search product, beyond the fact that responding to a query with one answer eliminates valuable real estate for online ads. A chatbot that pointed to one answer directly from Google could increase its liability if the response was found to be harmful or plagiarized.</div>
<div></div>
<div>Chatbots like OpenAI routinely make factual errors and often switch their answers depending on how a question is asked. Moving from providing a range of answers to queries that link directly to their source material, to using a chatbot to give a single, authoritative answer, would be a big shift that makes many inside Google nervous, said one former Google AI researcher. The company doesn’t want to take on the role or responsibility of providing single answers like that, the person said. Previous updates to search, such as adding Instant Answers, were done slowly and with great caution.</div>
<div></div>
<div>Inside Google, however, some of the frustration with the AI safety process came from the sense that cutting-edge technology was never released as a product because of fears of bad publicity — if, say, an AI model showed bias.</div>
<div>
<div></div>
<h2>How AI is changing culture</h2>
<div></div>
<div>How we make art</div>
<div>AI image generators are trained to “understand” the content of hundreds of millions of images, usually scraped from the internet (possibly including yours), in order to create new images out of thin air. Try one out for yourself.</div>
<div>1/3</div>
</div>
<div>Meta employees have also had to deal with the company’s concerns about bad PR, according to a person familiar with the company’s internal deliberations who spoke on the condition of anonymity to discuss internal conversations. Before launching new products or publishing research, Meta employees have to answer questions about the potential risks of publicizing their work, including how it could be misinterpreted, the person said. Some projects are reviewed by public relations staff, as well as internal compliance experts who ensure the company’s products comply with its 2011 Federal Trade Commission agreement on how it handles user data.</div>
<div></div>
<div>To Timnit Gebru, executive director of the nonprofit Distributed AI Research Institute, the prospect of Google sidelining its responsible AI team doesn’t necessarily signal a shift in power or safety concerns, because those warning of the potential harms were never empowered to begin with. “If we were lucky, we’d get invited to a meeting,” said Gebru, who helped lead Google’s Ethical AI team until she was fired for a paper criticizing large language models.</div>
<div></div>
<div>From Gebru’s perspective, Google was slow to release its AI tools because the company lacked a strong enough business incentive to risk a hit to its reputation.</div>
<div></div>
<div>Rumman Chowdhury, who led Twitter’s machine-learning ethics team until Elon Musk disbanded it in November, said she expects companies like Google to increasingly sideline internal critics and ethicists as they scramble to catch up with OpenAI.</div>
<div></div>
<div>“We thought it was going to be China pushing the U.S., but looks like it’s start-ups,” she said.</div>
<div></div>
<div><span class="gray-darkest" data-qa="attribution-text">By </span><a class="gray-darkest hover-gray-dark decoration-gray-dark underline hover-none decoration-1 underline-offset-1" href="https://www.washingtonpost.com/people/nitasha-tiku/" rel="author" data-qa="author-name">Nitasha Tiku</a>, <a class="gray-darkest hover-gray-dark decoration-gray-dark underline hover-none decoration-1 underline-offset-1" href="https://www.washingtonpost.com/people/gerrit-de-vynck/" rel="author" data-qa="author-name">Gerrit De Vynck</a> and <a class="gray-darkest hover-gray-dark decoration-gray-dark underline hover-none decoration-1 underline-offset-1" href="https://www.washingtonpost.com/people/will-oremus/" rel="author" data-qa="author-name">Will Oremus</a> <a href="https://www.washingtonpost.com/technology/2023/01/27/chatgpt-google-meta/" target="_blank" rel="noopener">source</a></div>
<div>
<hr />
</div>
<div></div>
<div>
<div class="sc-157agsr-1 sc-157agsr-2 iHvAmE">
<div class="sc-9tztzq-2 fkvsnH">
<header class="sc-1efpnfq-1 hQgdUv">
<h1 class="sc-1efpnfq-0 joZwQS">Amazon Warns Employees to Beware of ChatGPT</h1>
</header>
<h2 class="sc-1xcxnn7-0 iBCShr js_regular-subhead">At the same time, OpenAI&#8217;s Chat GPT gave correct answers to interview questions for a software coding position.</h2>
</div>
</div>
<div class="sc-157agsr-0 knYNlh">
<p>&nbsp;</p>
<div class="sc-11qwj9y-0 TCDpD">
<div class="js_starterpost">
<div class="sc-1fofo4n-0 cdfDtI">
<div class="sc-83o472-4 gxeLWy">
<div class="sc-83o472-0 cCffwz">
<div class="sc-1jc3ukb-0 eJFjvj">
<div class="sc-1jc3ukb-3 bPVaPw">By <a class="sc-1out364-0 hMndXN js_link" href="https://gizmodo.com/author/kevinhurler" data-ga="[[&quot;Permalink meta&quot;,&quot;Author click&quot;,&quot;https://kinja.com/kevinhurler&quot;]]">Kevin Hurler</a> <a href="https://gizmodo.com/amazon-chatgpt-ai-software-job-coding-1850034383" target="_blank" rel="noopener">source</a></div>
</div>
</div>
</div>
</div>
</div>
</div>
<p>&nbsp;</p>
</div>
</div>
<div></div>
<div>
<p class="sc-77igqf-0 bOfvBY">ChatGPT has been making the tech industry sweat since its rise in popularity last year, and now Amazon is feeling the heat too. According to internal communications from the company as viewed by Insider, an Amazon lawyer has urged employees not to share code with the AI chatbot.</p>
<div class="bxm4mm-2 hKBnez"></div>
<p class="sc-77igqf-0 bOfvBY">Insider reported earlier this week that the lawyer specifically requested that employees not share “any Amazon confidential information (including Amazon code you are working on)” with ChatGPT, according to screenshots of Slack messages reviewed by the outlet. The guidance comes after the company reportedly witnessed ChatGPT responses that have mimicked internal Amazon data.</p>
<div class="sc-1xxadal-0 jICWEe js_related-stories-inset">
<div class="sc-1xsbcvq-0 jpdpJT">
<h4 class="sc-1xxadal-1 ckNlOR">Related Stories</h4>
<div class="sc-1xxadal-2 bNjGjO">
<div class="sc-1xxadal-3 gGFGUT">Amazon UK Workers Stage First Official Fulfillment Center Strike</div>
<div class="sc-1xxadal-3 gGFGUT">Nobody’s Really Watching Reels</div>
<div class="sc-1xxadal-3 gGFGUT">Android Users Can Finally Use Alternate Search and Payment Methods&#8230; but Only in India</div>
</div>
</div>
</div>
<p class="sc-77igqf-0 bOfvBY">“This is important because your inputs may be used as training data for a further iteration of ChatGPT, and we wouldn’t want its output to include or resemble our confidential information (and I’ve already seen instances where its output closely matches existing material),” the lawyer wrote further, according to Insider.</p>
<p class="sc-77igqf-0 bOfvBY">Amazon is probably right about ChatGPT obtaining its data since, in a similar story, ChatGPT allegedly answered interview questions correctly for a software coding position at the company. According to Slack channel transcripts also reviewed by Insider, the AI provided correct solutions to software coding questions and even made improvements to some of Amazon’s code.</p>
<p class="sc-77igqf-0 bOfvBY">“I was honestly impressed!” an employee reportedly wrote on Slack. “I’m both scared and excited to see what impact this will have on the way that we conduct coding interviews.”</p>
<p class="sc-77igqf-0 bOfvBY">The novelty of ChatGPT has not warn off just yet, but questions surrounding how it may intersect with our daily lives have sprung up in recent weeks. While the chatbot was able to pass a final exam at Wharton in an MBA level course (where it did struggle with some basic arithmetic), ChatGPT’s role in education—among other fields—is dubious. Some school systems, like the New York City Department of Education, have decided to ban the tech over fears of cheating, but OpenAI’s CEO simply believes school administrators need to get over it.</p>
</div>
<div>
<div class="sc-157agsr-0 knYNlh">
<p>&nbsp;</p>
<div class="sc-11qwj9y-0 TCDpD">
<div class="js_starterpost">
<div class="sc-1fofo4n-0 cdfDtI">
<div class="sc-83o472-4 gxeLWy">
<div class="sc-83o472-0 cCffwz">
<div class="sc-1jc3ukb-2 fASvtw">
<div class="sc-1jc3ukb-1 eYTvqt"></div>
</div>
</div>
</div>
</div>
</div>
</div>
<p>&nbsp;</p>
</div>
</div>
<p>&nbsp;</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Fixing YouTube Search with OpenAI&#8217;s Whisper</title>
		<link>https://goodshepherdmedia.net/fixing-youtube-search-with-openais-whisper/</link>
		
		<dc:creator><![CDATA[The Truth News]]></dc:creator>
		<pubDate>Wed, 18 Jan 2023 10:27:41 +0000</pubDate>
				<category><![CDATA[Science & Engineering]]></category>
		<category><![CDATA[Tech]]></category>
		<category><![CDATA[Top Stories]]></category>
		<category><![CDATA[Zee Truthful News]]></category>
		<category><![CDATA[💻Tech History]]></category>
		<category><![CDATA[🤖 AI Artificial Intelligence]]></category>
		<category><![CDATA[🤖Open AI]]></category>
		<category><![CDATA[🤖🗣️Whisper]]></category>
		<category><![CDATA[Open AI]]></category>
		<category><![CDATA[Open AI's Whisper]]></category>
		<category><![CDATA[OpenAI's Whisper]]></category>
		<category><![CDATA[Whisper AI]]></category>
		<guid isPermaLink="false">https://goodshepherdmedia.net/?p=10212</guid>

					<description><![CDATA[Fixing YouTube Search with OpenAI&#8217;s Whisper OpenAI’s Whisper is a new state-of-the-art (SotA) model in speech-to-text. It is able to almost flawlessly transcribe speech across dozens of languages and even handle poor audio quality or excessive background noise. The domain of spoken word has always been somewhat out of reach for ML use-cases. Whisper changes that for [&#8230;]]]></description>
										<content:encoded><![CDATA[<h1>Fixing YouTube Search with OpenAI&#8217;s Whisper</h1>
<p>OpenAI’s <em>Whisper</em> is a new state-of-the-art (SotA) model in speech-to-text. It is able to almost flawlessly transcribe speech across dozens of languages and even handle poor audio quality or excessive background noise.</p>
<p>The domain of spoken word has always been somewhat out of reach for ML use-cases. Whisper changes that for speech-centric use cases. We will demonstrate the power of Whisper alongside other technologies like transformers and vector search by building a new and improved YouTube search.</p>
<p>Search on YouTube is good but has its limitations, especially when it comes to answering questions. With trillions of hours of content, there should be an answer to almost every question. Yet, if we have a specific question like <em>“what is OpenAI’s CLIP?&#8221;</em>, instead of a concise answer we get lots of very long videos that we must watch through.</p>
<p>What if all we want is a short 20-second explanation? The current YouTube search has no solution for this. Maybe there’s a good reason to encourage users to watch as much of a video as possible (more ads, anyone?).</p>
<p>Whisper is the solution to this problem <em>and many others involving the spoken word</em>. In this article, we’ll explore the idea behind a better speech-enabled search.</p>
<p><iframe title="How to Use OpenAI Whisper to Fix YouTube Search" width="640" height="360" src="https://www.youtube.com/embed/vpU_6x3jowg?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></p>
<p><iframe title="How to do Free Speech-to-Text Transcription Better Than Google Premium API with OpenAI Whisper Model" width="640" height="360" src="https://www.youtube.com/embed/msj3wuYf3d8?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></p>
<div></div>
<div></div>
<div></div>
<hr />
<h2 id="the-idea">The Idea</h2>
<p>We want to get specific timestamps that answer our search queries. YouTube does support time-specific links in videos, so a more precise search with these links should be possible.</p>
<p><video class="responsive" autoplay="autoplay" loop="loop" muted="" width="300" height="150" data-mce-fragment="1"></video></p>
<p><small>Timestamp URLs can be copied directly from a video, we can use the same URL format in our search app.</small></p>
<p>To build something like this, we first need to transcribe the audio in our videos to text. YouTube automatically captions every video, and the captions are okay — <em>but</em> OpenAI just open-sourced something called “Whisper”.</p>
<p>Whisper is best described as the GPT-3 or DALL-E 2 of speech-to-text. It’s open source and can transcribe audio in real-time <em>or faster</em> with <em>unparalleled performance</em>. That seems like the most exciting option.</p>
<p>Once we have our transcribed text and the timestamps for each text snippet, we can move on to the <a href="https://www.pinecone.io/learn/question-answering">question-answering (QA)</a> part. QA is a form of search where given a natural language query like <em>“what is OpenAI’s Whisper?&#8221;</em> we can return accurate natural language answers.</p>
<p>We can think of QA as the most intuitive form of searching for information because it is how we ask other people for information. The only difference being we type the question into a search bar rather than verbally communicate it — for now.</p>
<p>How does all of this look?</p>
<p><img fetchpriority="high" decoding="async" class="" src="https://d33wubrfki0l68.cloudfront.net/b04cbd7e64c2cbfe65bfe1f6b9035e239d845871/2c500/images/openai-whisper-2.png" alt="whisper-architecture" width="787" height="535" /></p>
<p><small>Overview of the process used in our demo. Covering OpenAI’s Whisper, sentence transformers, the Pinecone vector database, and more.</small></p>
<p>Now let’s color in the details and walk through the steps.</p>
<h2 id="video-data">Video Data</h2>
<p>The first step is to download our YouTube video data and extract the audio attached to each video. Fortunately, there’s a Python library for exactly that called <code>pytube</code>.</p>
<p>With <code>pytube</code>, we provide a video ID (found in the URL bar or downloadable if you have a channel). I directly downloaded a summary of channel content, including IDs, titles, publication dates, etc., via YouTube. This same data is available via Hugging Face <em>Datasets</em> in a dataset called <code>jamescalam/channel-metadata</code>.</p>
<div id="gist118847977" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-channel-meta-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="159f0d0f-0192-47a5-8a8b-817a070ea0f7" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=7fb608479778bb2a5220463f44241ee34a099638&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f62343238306431663430383935623066623130353832663332363732346332312f7261772f376662363038343739373738626232613532323034363366343432343165653334613039393633382f776869737065722d79742d7365617263682d6368616e6e656c2d6d6574612e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2Fb4280d1f40895b0fb10582f326724c21&amp;path=whisper-yt-search-channel-meta.ipynb&amp;repository_id=118847977&amp;repository_type=Gist#159f0d0f-0192-47a5-8a8b-817a070ea0f7" name="159f0d0f-0192-47a5-8a8b-817a070ea0f7" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/b4280d1f40895b0fb10582f326724c21/raw/7fb608479778bb2a5220463f44241ee34a099638/whisper-yt-search-channel-meta.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/b4280d1f40895b0fb10582f326724c21#file-whisper-yt-search-channel-meta-ipynb">whisper-yt-search-channel-meta.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>We’re most interested in the <code>Title</code> and <code>Video ID</code> fields. With the video ID, we can begin downloading the videos and saving the audio files with <code>pytube</code>.</p>
<div id="gist118847915" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-pytube-py" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-python  ">
<div class="js-check-bidi js-blob-code-container blob-code-content">
<table class="highlight tab-size js-file-line-container js-code-nav-container js-tagsearch-file" data-hpc="" data-tab-size="8" data-paste-markdown-skip="" data-tagsearch-lang="Python" data-tagsearch-path="whisper-yt-search-pytube.py">
<tbody>
<tr>
<td id="file-whisper-yt-search-pytube-py-L1" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="1"></td>
<td id="file-whisper-yt-search-pytube-py-LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-s1">pytube</span> <span class="pl-k">import</span> <span class="pl-v">YouTube</span> <span class="pl-c"># !pip install pytube</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L2" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="2"></td>
<td id="file-whisper-yt-search-pytube-py-LC2" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-s1">pytube</span>.<span class="pl-s1">exceptions</span> <span class="pl-k">import</span> <span class="pl-v">RegexMatchError</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L3" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="3"></td>
<td id="file-whisper-yt-search-pytube-py-LC3" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-s1">tqdm</span>.<span class="pl-s1">auto</span> <span class="pl-k">import</span> <span class="pl-s1">tqdm</span> <span class="pl-c"># !pip install tqdm</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L4" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="4"></td>
<td id="file-whisper-yt-search-pytube-py-LC4" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L5" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="5"></td>
<td id="file-whisper-yt-search-pytube-py-LC5" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># where to save</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L6" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="6"></td>
<td id="file-whisper-yt-search-pytube-py-LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">save_path</span> <span class="pl-c1">=</span> <span class="pl-s">&#8220;./mp3&#8221;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L7" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="7"></td>
<td id="file-whisper-yt-search-pytube-py-LC7" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L8" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="8"></td>
<td id="file-whisper-yt-search-pytube-py-LC8" class="blob-code blob-code-inner js-file-line"><span class="pl-k">for</span> <span class="pl-s1">i</span>, <span class="pl-s1">row</span> <span class="pl-c1">in</span> <span class="pl-en">tqdm</span>(<span class="pl-s1">videos_meta</span>):</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L9" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="9"></td>
<td id="file-whisper-yt-search-pytube-py-LC9" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># url of video to be downloaded</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L10" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="10"></td>
<td id="file-whisper-yt-search-pytube-py-LC10" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">url</span> <span class="pl-c1">=</span> <span class="pl-s">f&#8221;https://youtu.be/<span class="pl-s1"><span class="pl-kos">{</span>row[&#8216;Video ID&#8217;]<span class="pl-kos">}</span></span>&#8220;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L11" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="11"></td>
<td id="file-whisper-yt-search-pytube-py-LC11" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L12" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="12"></td>
<td id="file-whisper-yt-search-pytube-py-LC12" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># try to create a YouTube vid object</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L13" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="13"></td>
<td id="file-whisper-yt-search-pytube-py-LC13" class="blob-code blob-code-inner js-file-line"><span class="pl-k">try</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L14" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="14"></td>
<td id="file-whisper-yt-search-pytube-py-LC14" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">yt</span> <span class="pl-c1">=</span> <span class="pl-v">YouTube</span>(<span class="pl-s1">url</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L15" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="15"></td>
<td id="file-whisper-yt-search-pytube-py-LC15" class="blob-code blob-code-inner js-file-line"><span class="pl-k">except</span> <span class="pl-v">RegexMatchError</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L16" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="16"></td>
<td id="file-whisper-yt-search-pytube-py-LC16" class="blob-code blob-code-inner js-file-line"><span class="pl-en">print</span>(<span class="pl-s">f&#8221;RegexMatchError for &#8216;<span class="pl-s1"><span class="pl-kos">{</span>url<span class="pl-kos">}</span></span>&#8216;&#8221;</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L17" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="17"></td>
<td id="file-whisper-yt-search-pytube-py-LC17" class="blob-code blob-code-inner js-file-line"><span class="pl-k">continue</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L18" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="18"></td>
<td id="file-whisper-yt-search-pytube-py-LC18" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L19" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="19"></td>
<td id="file-whisper-yt-search-pytube-py-LC19" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">itag</span> <span class="pl-c1">=</span> <span class="pl-c1">None</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L20" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="20"></td>
<td id="file-whisper-yt-search-pytube-py-LC20" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># we only want audio files</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L21" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="21"></td>
<td id="file-whisper-yt-search-pytube-py-LC21" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">files</span> <span class="pl-c1">=</span> <span class="pl-s1">yt</span>.<span class="pl-s1">streams</span>.<span class="pl-en">filter</span>(<span class="pl-s1">only_audio</span><span class="pl-c1">=</span><span class="pl-c1">True</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L22" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="22"></td>
<td id="file-whisper-yt-search-pytube-py-LC22" class="blob-code blob-code-inner js-file-line"><span class="pl-k">for</span> <span class="pl-s1">file</span> <span class="pl-c1">in</span> <span class="pl-s1">files</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L23" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="23"></td>
<td id="file-whisper-yt-search-pytube-py-LC23" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># from audio files we grab the first audio for mp4 (eg mp3)</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L24" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="24"></td>
<td id="file-whisper-yt-search-pytube-py-LC24" class="blob-code blob-code-inner js-file-line"><span class="pl-k">if</span> <span class="pl-s1">file</span>.<span class="pl-s1">mime_type</span> <span class="pl-c1">==</span> <span class="pl-s">&#8216;audio/mp4&#8217;</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L25" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="25"></td>
<td id="file-whisper-yt-search-pytube-py-LC25" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">itag</span> <span class="pl-c1">=</span> <span class="pl-s1">file</span>.<span class="pl-s1">itag</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L26" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="26"></td>
<td id="file-whisper-yt-search-pytube-py-LC26" class="blob-code blob-code-inner js-file-line"><span class="pl-k">break</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L27" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="27"></td>
<td id="file-whisper-yt-search-pytube-py-LC27" class="blob-code blob-code-inner js-file-line"><span class="pl-k">if</span> <span class="pl-s1">itag</span> <span class="pl-c1">is</span> <span class="pl-c1">None</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L28" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="28"></td>
<td id="file-whisper-yt-search-pytube-py-LC28" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># just incase no MP3 audio is found (shouldn&#8217;t happen)</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L29" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="29"></td>
<td id="file-whisper-yt-search-pytube-py-LC29" class="blob-code blob-code-inner js-file-line"><span class="pl-en">print</span>(<span class="pl-s">&#8220;NO MP3 AUDIO FOUND&#8221;</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L30" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="30"></td>
<td id="file-whisper-yt-search-pytube-py-LC30" class="blob-code blob-code-inner js-file-line"><span class="pl-k">continue</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L31" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="31"></td>
<td id="file-whisper-yt-search-pytube-py-LC31" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L32" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="32"></td>
<td id="file-whisper-yt-search-pytube-py-LC32" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># get the correct mp3 &#8216;stream&#8217;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L33" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="33"></td>
<td id="file-whisper-yt-search-pytube-py-LC33" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">stream</span> <span class="pl-c1">=</span> <span class="pl-s1">yt</span>.<span class="pl-s1">streams</span>.<span class="pl-en">get_by_itag</span>(<span class="pl-s1">itag</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L34" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="34"></td>
<td id="file-whisper-yt-search-pytube-py-LC34" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># downloading the audio</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L35" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="35"></td>
<td id="file-whisper-yt-search-pytube-py-LC35" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">stream</span>.<span class="pl-en">download</span>(</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L36" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="36"></td>
<td id="file-whisper-yt-search-pytube-py-LC36" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">output_path</span><span class="pl-c1">=</span><span class="pl-s1">save_path</span>,</td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L37" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="37"></td>
<td id="file-whisper-yt-search-pytube-py-LC37" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">filename</span><span class="pl-c1">=</span><span class="pl-s">f&#8221;<span class="pl-s1"><span class="pl-kos">{</span>row[&#8216;Video ID&#8217;]<span class="pl-kos">}</span></span>.mp3&#8243;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-pytube-py-L38" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="38"></td>
<td id="file-whisper-yt-search-pytube-py-LC38" class="blob-code blob-code-inner js-file-line">)</td>
</tr>
</tbody>
</table>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/51c7d1cd1ad4e8c4b581f05c59881e39/raw/fe6d74835cb9cf1a0fe830ec0bf22f336c176555/whisper-yt-search-pytube.py">view raw</a><a href="https://gist.github.com/jamescalam/51c7d1cd1ad4e8c4b581f05c59881e39#file-whisper-yt-search-pytube-py">whisper-yt-search-pytube.py </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>After this, we should find ~108 audio MP3 files stored in the <code>./mp3</code> directory.</p>
<p><img decoding="async" class="" src="https://d33wubrfki0l68.cloudfront.net/9027b8e7c1bd21239f83d39c1c122da12f679524/8b071/images/openai-whisper-3.png" alt="mp3 files directory" width="1031" height="646" /></p>
<p><small>Downloaded MP3 files in the <code>./mp3</code> directory.</small></p>
<p>With these, we can move on to transcription with OpenAI’s Whisper.</p>
<h2 id="speech-to-text-with-whisper">Speech-to-Text with Whisper</h2>
<p>OpenAI’s Whisper speech-to-text-model is completely open source and available via <a href="https://github.com/openai/whisper">OpenAI’s Whisper library</a> available for <code>pip install</code> via GitHub:</p>
<pre><code>!pip install git+https://github.com/openai/whisper.git</code></pre>
<p>Whisper relies on another software called FFMPEG to convert video and audio files. The installation for this varies by OS [1]; the following cover the primary systems:</p>
<pre><code># on Ubuntu or Debian sudo apt update &amp;&amp; sudo apt install ffmpeg # on Arch Linux sudo pacman -S ffmpeg # on MacOS using Homebrew (https://brew.sh/) brew install ffmpeg # on Windows using Chocolatey (https://chocolatey.org/) choco install ffmpeg # on Windows using Scoop (https://scoop.sh/) scoop install ffmpeg</code></pre>
<p>After installation, we download and initialize the <em>large</em> model, moving it to GPU if CUDA is available.</p>
<div id="gist118848073" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-init-whisper-py" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-python  ">
<div class="js-check-bidi js-blob-code-container blob-code-content">
<table class="highlight tab-size js-file-line-container js-code-nav-container js-tagsearch-file" data-hpc="" data-tab-size="8" data-paste-markdown-skip="" data-tagsearch-lang="Python" data-tagsearch-path="whisper-yt-search-init-whisper.py">
<tbody>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L1" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="1"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> <span class="pl-s1">whisper</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L2" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="2"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC2" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> <span class="pl-s1">torch</span> <span class="pl-c"># install steps: pytorch.org</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L3" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="3"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC3" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L4" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="4"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC4" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">device</span> <span class="pl-c1">=</span> <span class="pl-s">&#8220;cuda&#8221;</span> <span class="pl-k">if</span> <span class="pl-s1">torch</span>.<span class="pl-s1">cuda</span>.<span class="pl-en">is_available</span>() <span class="pl-k">else</span> <span class="pl-s">&#8220;cpu&#8221;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L5" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="5"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC5" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-init-whisper-py-L6" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="6"></td>
<td id="file-whisper-yt-search-init-whisper-py-LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">model</span> <span class="pl-c1">=</span> <span class="pl-s1">whisper</span>.<span class="pl-en">load_model</span>(<span class="pl-s">&#8220;large&#8221;</span>).<span class="pl-en">to</span>(<span class="pl-s1">device</span>)</td>
</tr>
</tbody>
</table>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/9b41191c65c029602d85c08043f6f683/raw/f183e4344edd3e80e0a7f94ea5afbb103ee0f675/whisper-yt-search-init-whisper.py">view raw</a><a href="https://gist.github.com/jamescalam/9b41191c65c029602d85c08043f6f683#file-whisper-yt-search-init-whisper-py">whisper-yt-search-init-whisper.py </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>Other models are available, and given a smaller GPU (or even CPU) should be considered. We transcribe the audio like so:</p>
<div id="gist118848187" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-transcribe-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="4cab2014-8712-4bbb-b47c-93b6e1abda54" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=03b3d637baaad2ae9c0d294f87815d273d2e7b7c&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f38636439633131313230343539333864666565306539393534323232626331662f7261772f303362336436333762616161643261653963306432393466383738313564323733643265376237632f776869737065722d79742d7365617263682d7472616e7363726962652e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F8cd9c1112045938dfee0e9954222bc1f&amp;path=whisper-yt-search-transcribe.ipynb&amp;repository_id=118848187&amp;repository_type=Gist#4cab2014-8712-4bbb-b47c-93b6e1abda54" name="4cab2014-8712-4bbb-b47c-93b6e1abda54" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/8cd9c1112045938dfee0e9954222bc1f/raw/03b3d637baaad2ae9c0d294f87815d273d2e7b7c/whisper-yt-search-transcribe.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/8cd9c1112045938dfee0e9954222bc1f#file-whisper-yt-search-transcribe-ipynb">whisper-yt-search-transcribe.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>From this, we have a list of ~27K transcribed audio segments, including text alongside start and end seconds. If you are waiting a long time for this to process, a pre-built version of the dataset is available. Download instructions are in the following section.</p>
<p>The last cell from above is missing the logic required to extract and add the metadata from our <code>videos_dict</code> that we initialized earlier. We add that like so:</p>
<div id="gist118848525" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-build-segments-py" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-python  ">
<div class="js-check-bidi js-blob-code-container blob-code-content">
<table class="highlight tab-size js-file-line-container js-code-nav-container js-tagsearch-file" data-hpc="" data-tab-size="8" data-paste-markdown-skip="" data-tagsearch-lang="Python" data-tagsearch-path="whisper-yt-search-build-segments.py">
<tbody>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L1" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="1"></td>
<td id="file-whisper-yt-search-build-segments-py-LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">data</span> <span class="pl-c1">=</span> []</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L2" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="2"></td>
<td id="file-whisper-yt-search-build-segments-py-LC2" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L3" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="3"></td>
<td id="file-whisper-yt-search-build-segments-py-LC3" class="blob-code blob-code-inner js-file-line"><span class="pl-k">for</span> <span class="pl-s1">i</span>, <span class="pl-s1">path</span> <span class="pl-c1">in</span> <span class="pl-en">enumerate</span>(<span class="pl-en">tqdm</span>(<span class="pl-s1">paths</span>)):</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L4" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="4"></td>
<td id="file-whisper-yt-search-build-segments-py-LC4" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">_id</span> <span class="pl-c1">=</span> <span class="pl-s1">path</span>.<span class="pl-en">split</span>(<span class="pl-s">&#8216;/&#8217;</span>)[<span class="pl-c1">&#8211;</span><span class="pl-c1">1</span>][:<span class="pl-c1">&#8211;</span><span class="pl-c1">4</span>]</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L5" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="5"></td>
<td id="file-whisper-yt-search-build-segments-py-LC5" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># transcribe to get speech-to-text data</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L6" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="6"></td>
<td id="file-whisper-yt-search-build-segments-py-LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">result</span> <span class="pl-c1">=</span> <span class="pl-s1">model</span>.<span class="pl-en">transcribe</span>(<span class="pl-s1">path</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L7" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="7"></td>
<td id="file-whisper-yt-search-build-segments-py-LC7" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">segments</span> <span class="pl-c1">=</span> <span class="pl-s1">result</span>[<span class="pl-s">&#8216;segments&#8217;</span>]</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L8" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="8"></td>
<td id="file-whisper-yt-search-build-segments-py-LC8" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># get the video metadata&#8230;</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L9" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="9"></td>
<td id="file-whisper-yt-search-build-segments-py-LC9" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">video_meta</span> <span class="pl-c1">=</span> <span class="pl-s1">videos_dict</span>[<span class="pl-s1">_id</span>]</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L10" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="10"></td>
<td id="file-whisper-yt-search-build-segments-py-LC10" class="blob-code blob-code-inner js-file-line"><span class="pl-k">for</span> <span class="pl-s1">segment</span> <span class="pl-c1">in</span> <span class="pl-s1">segments</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L11" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="11"></td>
<td id="file-whisper-yt-search-build-segments-py-LC11" class="blob-code blob-code-inner js-file-line"><span class="pl-c"># merge segments data and videos_meta data</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L12" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="12"></td>
<td id="file-whisper-yt-search-build-segments-py-LC12" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">meta</span> <span class="pl-c1">=</span> {</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L13" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="13"></td>
<td id="file-whisper-yt-search-build-segments-py-LC13" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">**</span><span class="pl-s1">video_meta</span>,</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L14" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="14"></td>
<td id="file-whisper-yt-search-build-segments-py-LC14" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">**</span>{</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L15" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="15"></td>
<td id="file-whisper-yt-search-build-segments-py-LC15" class="blob-code blob-code-inner js-file-line"><span class="pl-s">&#8220;id&#8221;</span>: <span class="pl-s">f&#8221;<span class="pl-s1"><span class="pl-kos">{</span>_id<span class="pl-kos">}</span></span>-t<span class="pl-s1"><span class="pl-kos">{</span>segments[j][&#8216;start&#8217;]<span class="pl-kos">}</span></span>&#8220;</span>,</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L16" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="16"></td>
<td id="file-whisper-yt-search-build-segments-py-LC16" class="blob-code blob-code-inner js-file-line"><span class="pl-s">&#8220;text&#8221;</span>: <span class="pl-s1">segment</span>[<span class="pl-s">&#8220;text&#8221;</span>].<span class="pl-en">strip</span>(),</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L17" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="17"></td>
<td id="file-whisper-yt-search-build-segments-py-LC17" class="blob-code blob-code-inner js-file-line"><span class="pl-s">&#8220;start&#8221;</span>: <span class="pl-s1">segment</span>[<span class="pl-s">&#8216;start&#8217;</span>],</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L18" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="18"></td>
<td id="file-whisper-yt-search-build-segments-py-LC18" class="blob-code blob-code-inner js-file-line"><span class="pl-s">&#8220;end&#8221;</span>: <span class="pl-s1">segment</span>[<span class="pl-s">&#8216;end&#8217;</span>]</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L19" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="19"></td>
<td id="file-whisper-yt-search-build-segments-py-LC19" class="blob-code blob-code-inner js-file-line">}</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L20" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="20"></td>
<td id="file-whisper-yt-search-build-segments-py-LC20" class="blob-code blob-code-inner js-file-line">}</td>
</tr>
<tr>
<td id="file-whisper-yt-search-build-segments-py-L21" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="21"></td>
<td id="file-whisper-yt-search-build-segments-py-LC21" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">data</span>.<span class="pl-en">append</span>(<span class="pl-s1">meta</span>)</td>
</tr>
</tbody>
</table>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/4e6e978b5dcf7c4277d46f5f4a74798f/raw/7c18e20e8895e6cf46f11497a2a87e0799aad226/whisper-yt-search-build-segments.py">view raw</a><a href="https://gist.github.com/jamescalam/4e6e978b5dcf7c4277d46f5f4a74798f#file-whisper-yt-search-build-segments-py">whisper-yt-search-build-segments.py </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>After processing all of the segments, they are saved to file as a JSON lines file with:</p>
<div id="gist118848786" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-save-transcriptions-py" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-python  ">
<div class="js-check-bidi js-blob-code-container blob-code-content">
<table class="highlight tab-size js-file-line-container js-code-nav-container js-tagsearch-file" data-hpc="" data-tab-size="8" data-paste-markdown-skip="" data-tagsearch-lang="Python" data-tagsearch-path="whisper-yt-search-save-transcriptions.py">
<tbody>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L1" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="1"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> <span class="pl-s1">json</span></td>
</tr>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L2" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="2"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC2" class="blob-code blob-code-inner js-file-line"></td>
</tr>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L3" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="3"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC3" class="blob-code blob-code-inner js-file-line"><span class="pl-k">with</span> <span class="pl-en">open</span>(<span class="pl-s">&#8220;youtube-transcriptions.jsonl&#8221;</span>, <span class="pl-s">&#8220;w&#8221;</span>, <span class="pl-s1">encoding</span><span class="pl-c1">=</span><span class="pl-s">&#8220;utf-8&#8221;</span>) <span class="pl-k">as</span> <span class="pl-s1">fp</span>:</td>
</tr>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L4" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="4"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC4" class="blob-code blob-code-inner js-file-line"><span class="pl-k">for</span> <span class="pl-s1">line</span> <span class="pl-c1">in</span> <span class="pl-en">tqdm</span>(<span class="pl-s1">data</span>):</td>
</tr>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L5" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="5"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC5" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">json</span>.<span class="pl-en">dump</span>(<span class="pl-s1">line</span>, <span class="pl-s1">fp</span>)</td>
</tr>
<tr>
<td id="file-whisper-yt-search-save-transcriptions-py-L6" class="blob-num js-line-number js-code-nav-line-number js-blob-rnum" data-line-number="6"></td>
<td id="file-whisper-yt-search-save-transcriptions-py-LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-s1">fp</span>.<span class="pl-en">write</span>(<span class="pl-s">&#8216;<span class="pl-cce">\n</span>&#8216;</span>)</td>
</tr>
</tbody>
</table>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/e645ea4499e3e7984d4392325566573e/raw/6a0f2547e80e4fea7890719bb455ace3f7bc1d40/whisper-yt-search-save-transcriptions.py">view raw</a><a href="https://gist.github.com/jamescalam/e645ea4499e3e7984d4392325566573e#file-whisper-yt-search-save-transcriptions-py">whisper-yt-search-save-transcriptions.py </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>With that ready, let’s build the QA embeddings and vector search component.</p>
<h2 id="question-answering">Question-Answering</h2>
<p>On Hugging Face <em>Datasets</em>, you can find the data I scraped in a dataset called <code>jamescalam/youtube-transcriptions</code>:</p>
<div id="gist118848889" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-get-transcriptions-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="6605bf6a-ed33-423c-b81e-63895d87dff6" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=8fbb91a064a2898da35d4c095ca51a93487421f1&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f33653531376332656535666166666232626364626539663732376539643232372f7261772f386662623931613036346132383938646133356434633039356361353161393334383734323166312f776869737065722d79742d7365617263682d6765742d7472616e736372697074696f6e732e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F3e517c2ee5faffb2bcdbe9f727e9d227&amp;path=whisper-yt-search-get-transcriptions.ipynb&amp;repository_id=118848889&amp;repository_type=Gist#6605bf6a-ed33-423c-b81e-63895d87dff6" name="6605bf6a-ed33-423c-b81e-63895d87dff6" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/3e517c2ee5faffb2bcdbe9f727e9d227/raw/8fbb91a064a2898da35d4c095ca51a93487421f1/whisper-yt-search-get-transcriptions.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/3e517c2ee5faffb2bcdbe9f727e9d227#file-whisper-yt-search-get-transcriptions-ipynb">whisper-yt-search-get-transcriptions.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>For now, the dataset only contains videos from my personal channel, but I will add more videos from other ML-focused channels in the future.</p>
<p>The data includes a short chunk of text (the transcribed audio). Each chunk is relatively meaningless:</p>
<div id="gist118848859" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-short-segments-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="a43b2443-da86-43b9-bb97-c2d9bdface4d" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=5548ec17e93f0f54cb0b8b1f968709a3fa3cc597&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f64326562653032376336636438313236636166643362346632643338353235612f7261772f353534386563313765393366306635346362306238623166393638373039613366613363633539372f776869737065722d79742d7365617263682d73686f72742d7365676d656e74732e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2Fd2ebe027c6cd8126cafd3b4f2d38525a&amp;path=whisper-yt-search-short-segments.ipynb&amp;repository_id=118848859&amp;repository_type=Gist#a43b2443-da86-43b9-bb97-c2d9bdface4d" name="a43b2443-da86-43b9-bb97-c2d9bdface4d" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/d2ebe027c6cd8126cafd3b4f2d38525a/raw/5548ec17e93f0f54cb0b8b1f968709a3fa3cc597/whisper-yt-search-short-segments.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/d2ebe027c6cd8126cafd3b4f2d38525a#file-whisper-yt-search-short-segments-ipynb">whisper-yt-search-short-segments.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>Ideally, we want chunks of text 4-6x larger than this to capture enough meaning to be helpful. We do this by simply iterating over the dataset and merging every <em>six</em> segments.</p>
<div id="gist118848939" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-longer-segments-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="5ecd865a-1dd5-4a85-93ba-e6cf986307da" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=9ab91b16bf41870cc4949fd2107c8b8295aa03cb&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f31376466343031333364313163336332356161396634303435633964313134352f7261772f396162393162313662663431383730636334393439666432313037633862383239356161303363622f776869737065722d79742d7365617263682d6c6f6e6765722d7365676d656e74732e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F17df40133d11c3c25aa9f4045c9d1145&amp;path=whisper-yt-search-longer-segments.ipynb&amp;repository_id=118848939&amp;repository_type=Gist#5ecd865a-1dd5-4a85-93ba-e6cf986307da" name="5ecd865a-1dd5-4a85-93ba-e6cf986307da" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/17df40133d11c3c25aa9f4045c9d1145/raw/9ab91b16bf41870cc4949fd2107c8b8295aa03cb/whisper-yt-search-longer-segments.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/17df40133d11c3c25aa9f4045c9d1145#file-whisper-yt-search-longer-segments-ipynb">whisper-yt-search-longer-segments.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>A few things are happening here. First, we’re merging every six segments, as explained before. However, doing this alone will likely cut a lot of meaning between related segments.</p>
<p><img decoding="async" class="" src="https://d33wubrfki0l68.cloudfront.net/a0373a295dbbbe3c3ca686687b47a4e6c1aba11b/0c6f2/images/openai-whisper-4.png" alt="window-no-overlap" width="827" height="364" /></p>
<p><small>Even when merging segments we’re still left with a point where we must split the text (annotated with red cross-mark above). This can lead to us missing important information.</small></p>
<p>A common technique to avoid cutting related segments is adding some <em>overlap</em> between segments, where <code>stride</code> is used. For each step, we move <em>three</em> segments forward while merging <em>six</em> segments. By doing this, any meaningful segments cut in one step will be included in the next.</p>
<p><img loading="lazy" decoding="async" class="" src="https://d33wubrfki0l68.cloudfront.net/06f2cfe89c666aeddb1204b9741ae8a964460fb3/f6190/images/openai-whisper-5.png" alt="window-overlap" width="829" height="362" /></p>
<p><small>We can avoid this loss of meaning by adding an overlap when merging segments. It returns more data but means we are much less likely to cut between meaning segments.</small></p>
<p>With this, we have larger and more meaningful chunks of text. Now we need to encode them with a QA embedding model. Many high-performing, pretrained QA models are available via Hugging Face <em>Transformers</em> and the <em>Sentence Transformers</em> library. We will use one called <a href="https://huggingface.co/sentence-transformers/multi-qa-mpnet-base-dot-v1"><code>multi-qa-mpnet-base-dot-v1</code></a>.</p>
<div id="gist118848976" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-init-encoder-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="2e2ede93-df7b-44ce-b037-8355f0809b1b" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=7b8b0b3094761e66ebf905ddf933be9872f644f6&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f30643266616337326636643630613339666666343737613163326364663362612f7261772f376238623062333039343736316536366562663930356464663933336265393837326636343466362f776869737065722d79742d7365617263682d696e69742d656e636f6465722e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F0d2fac72f6d60a39fff477a1c2cdf3ba&amp;path=whisper-yt-search-init-encoder.ipynb&amp;repository_id=118848976&amp;repository_type=Gist#2e2ede93-df7b-44ce-b037-8355f0809b1b" name="2e2ede93-df7b-44ce-b037-8355f0809b1b" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/0d2fac72f6d60a39fff477a1c2cdf3ba/raw/7b8b0b3094761e66ebf905ddf933be9872f644f6/whisper-yt-search-init-encoder.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/0d2fac72f6d60a39fff477a1c2cdf3ba#file-whisper-yt-search-init-encoder-ipynb">whisper-yt-search-init-encoder.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>Using this model, we can encode a passage of text to a <em>meaningful</em> 768-dimensional vector with <code>model.encode("&lt;some text&gt;")</code>. Encoding all of our segments at once or storing them locally would require too much compute or memory — so we first initialize the vector database where they will be stored:</p>
<div id="gist118849004" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-init-pinecone-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="38e5e5c8-4247-42f6-946a-56879b89ede5" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=a11d7c1d105515aa0327dcc0a42fed4c2d7b77f2&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f65656261613262643435346635383237623736343330386137313437373436632f7261772f613131643763316431303535313561613033323764636330613432666564346332643762373766322f776869737065722d79742d7365617263682d696e69742d70696e65636f6e652e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2Feebaa2bd454f5827b764308a7147746c&amp;path=whisper-yt-search-init-pinecone.ipynb&amp;repository_id=118849004&amp;repository_type=Gist#38e5e5c8-4247-42f6-946a-56879b89ede5" name="38e5e5c8-4247-42f6-946a-56879b89ede5" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/eebaa2bd454f5827b764308a7147746c/raw/a11d7c1d105515aa0327dcc0a42fed4c2d7b77f2/whisper-yt-search-init-pinecone.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/eebaa2bd454f5827b764308a7147746c#file-whisper-yt-search-init-pinecone-ipynb">whisper-yt-search-init-pinecone.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>We should see that the index (vector database) is currently empty with a <code>total_vector_count</code> of <code>0</code>. Now we can begin encoding our segments and inserting the embeddings (and metadata) into our index.</p>
<div id="gist118849042" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-index-vecs-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="e0517d3d-2c66-455a-b588-75842ec40888" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=c582150a54ba8fe07d9e5e4b702e730b070b9bd2&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f34386134316239363663303634646164656431663638636561393164336638362f7261772f633538323135306135346261386665303764396535653462373032653733306230373062396264322f776869737065722d79742d7365617263682d696e6465782d766563732e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F48a41b966c064daded1f68cea91d3f86&amp;path=whisper-yt-search-index-vecs.ipynb&amp;repository_id=118849042&amp;repository_type=Gist#e0517d3d-2c66-455a-b588-75842ec40888" name="e0517d3d-2c66-455a-b588-75842ec40888" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/48a41b966c064daded1f68cea91d3f86/raw/c582150a54ba8fe07d9e5e4b702e730b070b9bd2/whisper-yt-search-index-vecs.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/48a41b966c064daded1f68cea91d3f86#file-whisper-yt-search-index-vecs-ipynb">whisper-yt-search-index-vecs.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>That is everything we needed to prepare our data and add everything to the vector database. All that is left is querying and returning results.</p>
<h2 id="making-queries">Making Queries</h2>
<p>Queries are straightforward to make; we:</p>
<ol>
<li>Encode the query using the same embedding model we used to encode the segments.</li>
<li>Pass to query to our index.</li>
</ol>
<p>We do that with the following:</p>
<div id="gist118849355" class="gist">
<div class="gist-file" translate="no">
<div class="gist-data">
<div class="js-gist-file-update-container js-task-list-container file-box">
<div id="file-whisper-yt-search-query-ipynb" class="file my-2">
<div class="Box-body p-0 blob-wrapper data type-jupyter-notebook  ">
<div class="render-wrapper ">
<div class="render-container is-render-pending js-render-target " data-identity="2c15dc64-c8b8-45f6-bfe4-c3cb9423181e" data-host="https://notebooks.githubusercontent.com" data-type="ipynb"><iframe class="render-viewer " title="File display" src="https://notebooks.githubusercontent.com/view/ipynb?color_mode=auto&amp;commit=2a439b5f3964c0d504b2478b7477b2ce247377bc&amp;enc_url=68747470733a2f2f7261772e67697468756275736572636f6e74656e742e636f6d2f676973742f6a616d657363616c616d2f32396235656564346135303338636436363533313135663434633332353761322f7261772f326134333962356633393634633064353034623234373862373437376232636532343733373762632f776869737065722d79742d7365617263682d71756572792e6970796e62&amp;logged_in=false&amp;nwo=jamescalam%2F29b5eed4a5038cd6653115f44c3257a2&amp;path=whisper-yt-search-query.ipynb&amp;repository_id=118849355&amp;repository_type=Gist#2c15dc64-c8b8-45f6-bfe4-c3cb9423181e" name="2c15dc64-c8b8-45f6-bfe4-c3cb9423181e" sandbox="allow-scripts allow-same-origin allow-top-navigation" data-mce-fragment="1"></iframe></div>
</div>
</div>
</div>
</div>
</div>
<div class="gist-meta"><a href="https://gist.github.com/jamescalam/29b5eed4a5038cd6653115f44c3257a2/raw/2a439b5f3964c0d504b2478b7477b2ce247377bc/whisper-yt-search-query.ipynb">view raw</a><a href="https://gist.github.com/jamescalam/29b5eed4a5038cd6653115f44c3257a2#file-whisper-yt-search-query-ipynb">whisper-yt-search-query.ipynb </a>hosted with <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/2764.png" alt="❤" class="wp-smiley" style="height: 1em; max-height: 1em;" /> by <a href="https://github.com/">GitHub</a></div>
</div>
</div>
<p>These results are relevant to the question; three, in particular, are from a similar location in the same video. We might want to improve the search interface to be more user-friendly than a Jupyter Notebook.</p>
<p>One of the easiest ways to get a web-based search UI up and running is with Hugging Face <em>Spaces</em> and Streamlit (or Gradio if preferred).</p>
<p>We won’t go through the code here, but if you’re familiar with Streamlit, you can build a search app quite easily within a few hours. Or you can use our example and do it in 5-10 minutes.</p>
<p>When querying again for <code>"what is OpenAI's clip?"</code> we can see that multiple results from a single video are merged. With this, we can jump to each segment by clicking on the part of the text that is most interesting to us.</p>
<p>Try a few more queries like:</p>
<pre><code>What is the best unsupervised method to train a sentence transformer? What is vector search? How can I train a sentence transformer with little-to-no data?</code></pre>
<hr />
<p>We can build incredible speech-enabled search apps very quickly using Whisper alongside Hugging Face, sentence transformers, and Pinecone’s <a href="https://www.pinecone.io/learn/vector-database">vector database</a>.</p>
<p>Whisper has unlocked a entire modality — the spoken word — and it’s only a matter of time before we see a significant increase in speech-enabled search and other speech-centric use cases.</p>
<p>Both machine learning and vector search have seen exponential growth in the past years. These technologies already seem like sci-fi. Yet, despite the incredible performance of everything we used here, it’s only a matter of time before all of this gets <em>even better</em>.</p>
<p><a href="https://www.pinecone.io/learn/openai-whisper/" target="_blank" rel="noopener">source</a></p>
<p><iframe title="How to do Free Speech-to-Text Transcription Better Than Google Premium API with OpenAI Whisper Model" width="640" height="360" src="https://www.youtube.com/embed/msj3wuYf3d8?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What is Chat GPT? What is Chat GPT Potential Uses</title>
		<link>https://goodshepherdmedia.net/what-is-chat-gpt-what-is-chat-gpt-potential-uses/</link>
		
		<dc:creator><![CDATA[The Truth News]]></dc:creator>
		<pubDate>Sun, 01 Jan 2023 08:28:42 +0000</pubDate>
				<category><![CDATA[Science & Engineering]]></category>
		<category><![CDATA[Tech]]></category>
		<category><![CDATA[Top Stories]]></category>
		<category><![CDATA[Zee Truthful News]]></category>
		<category><![CDATA[🤖 AI Artificial Intelligence]]></category>
		<category><![CDATA[🤖Open AI]]></category>
		<category><![CDATA[🤖🗣️ChatGPT]]></category>
		<category><![CDATA[Chat GPT]]></category>
		<category><![CDATA[Chat GPT-3]]></category>
		<category><![CDATA[ChatGPT]]></category>
		<category><![CDATA[Elon]]></category>
		<category><![CDATA[Elon Musk]]></category>
		<category><![CDATA[Musk]]></category>
		<category><![CDATA[Open AI]]></category>
		<category><![CDATA[Open AI Chat GPT-3]]></category>
		<guid isPermaLink="false">https://goodshepherdmedia.net/?p=10153</guid>

					<description><![CDATA[What is Chat GPT? What is Chat GPT Potential Uses Chat GPT Potential Use Cases &#160; How Chat GPT Can Solve Real World Problems! The latest tech product by Elon Musk founded Open AI is Chat GPT and it is nothing short of extraordinary. There hasn&#8217;t been a single day I did not use it [&#8230;]]]></description>
										<content:encoded><![CDATA[<blockquote>
<h1>What is Chat GPT? What is Chat GPT Potential Uses</h1>
<h2 style="text-align: center;"><span style="color: #0000ff;"><em>Chat GPT Potential Use Cases</em></span></h2>
</blockquote>
<p>&nbsp;</p>
<h2 class="title is-4 has-text-weight-bold mb-1" style="text-align: center;">How Chat GPT Can Solve Real World Problems!</h2>
<p>The latest tech product by Elon Musk founded Open AI is Chat GPT and it is nothing short of extraordinary. There hasn&#8217;t been a single day I did not use it since its release to the public on 1 Dec. GPT-3 (short for &#8220;Generative Pretrained Transformer 3&#8221;) is a state-of-the-art language processing model developed by OpenAI. It has a wide range of potential use cases, including natural language generation, translation, summarization, text classification, and more.</p>
<p>Chat GPT is one of its applications that is optimized for human-like conversations. When you give a prompt, the AI will reply like a human would do. You can even train it with your commands and change its tone &amp; style of chatting. No wonder it will help you save a ton of time and effort in researching topics on google. The possibilities are endless.</p>
<p>In this blog, we are going to explore some of the possible use cases of Chat GPT that can solve real-world problems.</p>
<p>&nbsp;</p>
<h3>Chat GPT Potential Use Cases</h3>
<ol>
<li><strong>Chatbot Development:</strong> Generate human-like text, a chatbot powered by GPT-3 could provide more natural and engaging interactions with users.</li>
<li><strong>Content Creation:</strong> Generate unique and compelling articles, blog posts, or social media content, saving time and effort for content marketers and writers.</li>
<li><strong>Virtual Assistants &amp; Digital Assistants:</strong> Generate responses to user queries, a virtual assistant could provide more accurate and natural-sounding answers to a wider range of questions.</li>
<li><strong>Translation:</strong> Translate any length of text into any language that AI is trained on. You don&#8217;t need to learn multiple languages to communicate.</li>
<li><strong>Summarization:</strong> Summarize a book or a text script or any long piece. Even create different interpretations of the same text.</li>
<li><strong>Question Answering:</strong> Ask a question and get the answer. This has applications in many Edtech products. It can even provide an explanation behind the answer.</li>
<li><strong>Social Media Moderation:</strong> Detect and flag inappropriate or offensive content on social media automatically.</li>
<li><strong>Sentence Completion &amp; Paraphrasing:</strong> Complete or paraphrase sentences of any length. Even decide the tone &amp; style of the paraphrased text.</li>
<li><strong>Text-to-speech and speech-to-text:</strong> Convert text to speech or speech to text.</li>
<li><strong>Named Entity Recognition:</strong> Automatically identify and classify named entities (people, places, organizations, etc.) in text.</li>
<li><strong>Write Novels, Stories &amp; Scripts:</strong> Develop a story or a film script with different scenes, plots, characters, dialogues etc. Creative storytelling with realistic and engaging elements.</li>
</ol>
<h3>Conclusion</h3>
<p>This language-processing AI model is so powerful that it will change the whole world. Chat GPT has the potential to revolutionize the way we interact with technology and each other. From providing personalized customer service to enhancing language learning, the potential use cases for Chat GPT are numerous and varied. As technology continues to evolve and improve, we can expect to see more and more creative and innovative applications for Chat GPT in the future. It&#8217;s an exciting time for this technology, and I can&#8217;t wait to see what the future holds. What about you?</p>
<p><a class="nuxt-link-active" href="https://leofinance.io/@finguru" target="_blank" rel="noopener" data-v-4b9f4c1a=""><img loading="lazy" decoding="async" class="is-rounded" src="https://images.hive.blog/u/finguru/avatar/small" width="45" height="45" data-v-30d64927="" data-v-4b9f4c1a="" /></a><a href="https://leofinance.io/@finguru/how-chat-gpt-can-solve-real-world-problems" target="_blank" rel="noopener">source</a> <a href="https://leofinance.io/@finguru" target="_blank" rel="noopener">@finguru</a></p>
<hr />
<div class="is-flex is-justify-content-start mr-2" data-v-55ff7877="">
<div class="ml-2 is-flex is-flex-direction-column" data-v-55ff7877="">
<div class="is-size-7 has-tooltip" data-v-55ff7877="" data-original-title="null">
<h1 class="fs-headline speakable-headline font-base font-size should-redesign" style="text-align: center;">12 Ways Chat GPT Will Change U.S. Forever</h1>
<blockquote>
<p style="text-align: center;"><strong><em><span style="color: #ff00ff;">Even Touching Areas Such as Immigration</span></em></strong></p>
</blockquote>
<figure class="embed-base image-embed embed-0" role="presentation">
<div><img loading="lazy" decoding="async" class="alignnone size-full wp-image-10161" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/960x0.jpg" alt="" width="960" height="422" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/960x0.jpg 960w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/960x0-300x132.jpg 300w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/960x0-768x338.jpg 768w" sizes="(max-width: 960px) 100vw, 960px" /></div><figcaption>
<p class="color-body light-text" aria-expanded="false">Chat CPT is a new technology that could change our world. It holds great promise to make our lives <span class="plus" data-ga-track="caption expand">&#8230;</span>GETTY</p>
</figcaption></figure>
<p>On November 30th, 2022 OpenAI, an American artificial intelligence research laboratory with a mission of developing friendly AI in a way that benefits humanity as a whole, came out with its cutting-edge Chat GPT chatbox. If Chat CPT lives up to its promise it will cause a tectonic shift in our lives. The appearance of Chat GPT is said to rival Google in its utility and has taken the world by storm attracting over 1 million users in just five days. Chat GPT appears ready to turn things upside down by impacting various occupations such as law, journalism, translations, and social media marketing. It works like Google in that you type a question into a box and hit enter. The chatbox displays the answer, usually in essay format, responding to the question posed. Let’s take a look at how Chat GPT can impact U.S. immigration as an example of what can be expected.</p>
<p>Immigrants might ask Chat GPT to outline what kinds of immigration options they have for immigrating to the United States. Then, they might ask to present the answer in a table format. Then applicants might ask Chat GPT to detail the requirements for a specific work visa and what will need to be included in an application. Chat GPT will respond as requested in real-time.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">2. Explaining Legislation</h3>
<p>Applicants for immigration benefits will be able to ask Chat GPT to explain immigration legislation in a simplified form. For example, investor immigrants might ask Chat GPT to explain the recent U.S. EB-5 Reform and Integrity Act of 2022 to better understand what is required of them in this manner. The simplified resulting version will make life much easier for the viewer who is spared the agony of reading the densely worded act and explanations.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">3. Specialized Instructions</h3>
<p>Immigration applicants will be able to ask Chat GPT to provide them with a specialized instruction guide on how they should prepare their immigrant applications to ensure the most likely approval of their cases. For example, this might include a request for a “Possible Table of Contents for an introduction to the U.S. E-2 work visa.” Chat GPT will type out the table. Then the applicant will be able to study up on the visa by typing in the various subtitles or dig deeper into any particular aspect of the visa, such as the quickest way to get one.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">4. Preparing Persuasive Resumes</h3>
<p>Chat GPT will be able to help applicants prepare persuasive resumes, for example by listing skills, education, experience, and projects undertaken by gathering the information from online sources or from online letters or text provided in written format to Chat GPT. Users will be able to type in requirements for positions and ask Chat GPT to suggest formats for resumes to match the requirements listed better.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">5. Preparing Covering Letters</h3>
<p>Chat GPT will be able to help applicants prepare cover letters drafted in a way that will reference their geographic locations, or expertise, or other items matching them up with the purpose of the letters.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">6. Doing Business Plans</h3>
<p>Chat GPT will be to help investor immigrant applicants prepare business plans for their proposed investment. For example, an immigrant investor proposing to buy a Holiday Inn Hotel in North Carolina will be able to ask Chat GPT to prepare the plan to apply for an E-2 visa for such a hotel using information published about Holiday Inns online.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">7. Helping To Learn English</h3>
<p>Chat GPT will be able to help applicants prepare for English and other language tests, such as the International English Language Testing System (IELTS) by providing examples of questions previously asked and answered. Indeed, the chatbox will be useful for learning new languages. For example, the chatbox could be asked, “What is the quickest way to learn the conjugations of verbs in Spanish?”</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">8. Preparing Country Profiles For Asylum Claimants</h3>
<p>Chat GPT will be available to asylum claimants to prepare detailed country profile information to strengthen submissions related to their claim of having a well-founded fear of persecution in escaping from their home country.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">9. Preparing Histories Of Events</h3>
<p>Government adjudicators of immigration applications will be able to use Chat GPT to prepare a complete history of events to verify the truth of answers to application forms and immigration submissions. Similarly, in depositions, attorneys will be able to search out histories of events to test the veracity of their evidence. But Chat GPT will not provide personal information such as location, age, or habits of indiviudals.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">10. Finding The Whereabouts Of Places</h3>
<p>Chat GPT will help users search to find places and their histories. But it will not be able to locate backgrounds such as of lost relatives who could sponsor users.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">11. Legal Research</h3>
<p>Chat GPT can search for key precedents, provide model pleadings and outline legal arguments for cases headed to immigration courts or the federal courts.</p>
<h3 class="subhead3-embed color-body bg-base font-accent font-size text-align">12. Translations</h3>
<p>Chat GPT will be able to translate documents for submission with immigration applications.</p>
<p>Chat GPT is still in beta mode and is offered for free to collect feedback on how to improve it. Its input is only up to date to 2021 and therefore it is not able to answer questions about current affairs or make projections about the future. Eventually, it will be made available commercially. You can try it <a class="color-link" title="https://chat.openai.com/chat" href="https://chat.openai.com/chat" target="_blank" rel="nofollow noopener noreferrer" data-ga-track="ExternalLink:https://chat.openai.com/chat" aria-label="here">here</a>. It’s great!</p>
<p>To conclude here is a copy of a question and answer I posed to Chat GPT about this article:</p>
<h2 class="subhead-embed color-accent bg-base font-accent font-size text-align">What outline would be best to introduce the benefits of Chat GPT to immigrants?</h2>
<p>Here is the actual Chat GPT answer:</p>
<p>Here is an outline that you could use to introduce the benefits of Chat GPT to immigrants:</p>
<ol>
<li>Introduction: Begin by explaining what Chat GPT is and why it might be useful for immigrants. You could mention that Chat GPT is a language model that can generate human-like text, and that it could potentially be used to assist with tasks such as answering frequently asked questions or generating documents.</li>
<li>Benefits: Next, outline the benefits of using Chat GPT for immigrants. For example, you could mention that Chat GPT could help immigrants save time by automating certain tasks, or that it could help them access information more easily. You could also mention that Chat GPT could potentially help immigrants navigate complex legal processes more effectively.</li>
<li>Examples: Provide some specific examples of how Chat GPT could be used by immigrants. For example, you could mention that it could be used to generate personalized responses to questions about the immigration process, or to generate documents such as visas or green card applications.</li>
<li>Limitations: It&#8217;s important to be upfront about the limitations of Chat GPT as well. For example, you could mention that it is not a substitute for legal advice, and that immigrants should always review any documents generated by Chat GPT before submitting them.</li>
<li>Conclusion: Summarize the main points of your introduction to Chat GPT, and encourage immigrants to consider using it as a tool to assist with their immigration needs.</li>
</ol>
<p><a class="contrib-link--name remove-underline not-premium-contrib-link--name" href="https://www.forbes.com/sites/andyjsemotiuk/">Andy J. Semotiuk</a> <a href="https://www.forbes.com/sites/andyjsemotiuk/2022/12/16/12-ways-chat-gpt-will-change-us-immigration-law-forever/?sh=2fc2562f6a5e" target="_blank" rel="noopener">source</a></p>
<hr />
<h1 class="story-title" style="text-align: center;">Chat GPT 3: A Look At The Artificial Intelligence Backed Chat Bot</h1>
<h3 class="story-description" style="text-align: center;">If you are on Twitter, even as a lurker and not an active user,<br />
it is quite likely you have come across Chat GPT 3. The free research preview of it is out now.</h3>
</div>
</div>
</div>
<p>If you are on Twitter, even as a lurker and not an active user, it is quite likely you have come across Chat GPT 3. The free research preview of it is out now. Chat GPT 3 made quite the impression on Twitter, just like the <a href="https://www.republicworld.com/world-news/us-news/researchers-ai-in-connected-cars-eased-rush-hour-congestion-articleshow.html">AI</a> art-generating app Midjourney did.</p>
<p>There were talks about how the GPT 3 will replace Google. Chamath Palihapitiya, a venture capitalist who worked at Facebook during its initial days, talked about how the Chat GPT 3 was much better than Google search. However, there are some problems. For starters &#8211; Chat GPT 3&#8217;s training data ends in the year 2021. As a result, it is completely unaware of anything that will happen in the world after 2021. This means, if we want to search something about the Russia-Ukraine war, then Google will be the go-to destination.</p>
<div id="ds_anchor_desktop"></div>
<h2>Initial impressions of Chat GPT 3</h2>
<p>&#8220;As I just learned that OpenAI had access to Twitter database for training. I put that on pause for now,&#8221; Musk tweeted, as people posted screenshots of their chat with Chat GPT 3. Chat GPT 3 &#8220;was started as open-source and non-profit. Neither are still true,&#8221; he added. Marc Andreessen of the venture capital firm Andreessen and Horowitz, has complained about bias in the responses of Chat GPT 3. Some users on Twitter have also complained about its inability to do basic linear algebra.</p>
<div data-oembed-url="https://mobile.twitter.com/pmarca/status/1605746351349735425">
<div class="twitter-tweet twitter-tweet-rendered"><iframe id="twitter-widget-0" class="" title="Twitter Tweet" src="https://platform.twitter.com/embed/Tweet.html?creatorScreenName=republic&amp;dnt=true&amp;embedId=twitter-widget-0&amp;features=eyJ0ZndfdGltZWxpbmVfbGlzdCI6eyJidWNrZXQiOltdLCJ2ZXJzaW9uIjpudWxsfSwidGZ3X2ZvbGxvd2VyX2NvdW50X3N1bnNldCI6eyJidWNrZXQiOnRydWUsInZlcnNpb24iOm51bGx9LCJ0ZndfdHdlZXRfZWRpdF9iYWNrZW5kIjp7ImJ1Y2tldCI6Im9uIiwidmVyc2lvbiI6bnVsbH0sInRmd19yZWZzcmNfc2Vzc2lvbiI6eyJidWNrZXQiOiJvbiIsInZlcnNpb24iOm51bGx9LCJ0Zndfc2hvd19idXNpbmVzc192ZXJpZmllZF9iYWRnZSI6eyJidWNrZXQiOiJvbiIsInZlcnNpb24iOm51bGx9LCJ0ZndfbWl4ZWRfbWVkaWFfMTU4OTciOnsiYnVja2V0IjoidHJlYXRtZW50IiwidmVyc2lvbiI6bnVsbH0sInRmd19leHBlcmltZW50c19jb29raWVfZXhwaXJhdGlvbiI6eyJidWNrZXQiOjEyMDk2MDAsInZlcnNpb24iOm51bGx9LCJ0ZndfZHVwbGljYXRlX3NjcmliZXNfdG9fc2V0dGluZ3MiOnsiYnVja2V0Ijoib24iLCJ2ZXJzaW9uIjpudWxsfSwidGZ3X3ZpZGVvX2hsc19keW5hbWljX21hbmlmZXN0c18xNTA4MiI6eyJidWNrZXQiOiJ0cnVlX2JpdHJhdGUiLCJ2ZXJzaW9uIjpudWxsfSwidGZ3X3Nob3dfYmx1ZV92ZXJpZmllZF9iYWRnZSI6eyJidWNrZXQiOiJvbiIsInZlcnNpb24iOm51bGx9LCJ0ZndfbGVnYWN5X3RpbWVsaW5lX3N1bnNldCI6eyJidWNrZXQiOnRydWUsInZlcnNpb24iOm51bGx9LCJ0Zndfc2hvd19nb3ZfdmVyaWZpZWRfYmFkZ2UiOnsiYnVja2V0Ijoib24iLCJ2ZXJzaW9uIjpudWxsfSwidGZ3X3Nob3dfYnVzaW5lc3NfYWZmaWxpYXRlX2JhZGdlIjp7ImJ1Y2tldCI6Im9uIiwidmVyc2lvbiI6bnVsbH0sInRmd190d2VldF9lZGl0X2Zyb250ZW5kIjp7ImJ1Y2tldCI6Im9uIiwidmVyc2lvbiI6bnVsbH19&amp;frame=false&amp;hideCard=false&amp;hideThread=false&amp;id=1605746351349735425&amp;lang=en&amp;origin=https%3A%2F%2Fwww.republicworld.com%2Ftechnology-news%2Fother-tech-news%2Fchat-gpt-3-a-look-at-the-artificial-intelligence-backed-chat-bot-articleshow.html&amp;sessionId=d8caab9fc4bfdd3349f9671416cbb2ccdb5cf74b&amp;siteScreenName=republic&amp;theme=light&amp;widgetsVersion=aaf4084522e3a%3A1674595607486&amp;width=550px" frameborder="0" scrolling="no" allowfullscreen="allowfullscreen" data-tweet-id="1605746351349735425" data-mce-fragment="1"></iframe></div>
</div>
<p>Chat GPT 3 can do some things that Google can&#8217;t. For example, Google can&#8217;t write a play. Chat GPT 3 on the other hand can. If you use it to write multiple plays, a pattern can be detected and the plays start becoming monotonous, but when one commands Chat GPT 3 to write a play for the first time, the result can be interesting, if not perfect. One merely has to give it a few prompts and voila, a play is there.</p>
<h2>An example of Chat GPT 3&#8217;s work</h2>
<blockquote><p>Consider this prompt &#8211; &#8220;Write a drama on a horror story set in a British University. Characters &#8211; Professor James; post graduate student Emily; Undergrad student Mark; Emily&#8217;s neighbour William&#8221;.</p>
<p>The result of the prompt was this &#8211; &#8221;</p>
<p>Professor James sat at his desk, surrounded by piles of papers and old books. He was a well-respected scholar of ancient history at the University of Oxford, and he had been working on his latest research project for months.</p>
<p>As he sorted through his notes, there was a knock at the door. &#8220;Come in,&#8221; he called out, not looking up from his work.</p>
<p>The door opened, and a young woman named Emily walked in. She was a postgraduate student, and she had been working with Professor James on his research.</p>
<p>&#8220;Hello, Professor,&#8221; Emily said, taking a seat across from him. &#8220;I&#8217;ve been going through some of the old documents you gave me, and I think I&#8217;ve found something interesting.&#8221;</p>
<p>Professor James looked up, intrigued. &#8220;What have you found?&#8221; he asked.</p>
<p>&#8220;It&#8217;s a letter from the 18th century,&#8221; Emily explained. &#8220;It&#8217;s written in Latin, and it mentions a mysterious ritual that was performed at the university. It&#8217;s supposed to be able to bring the dead back to life.&#8221;</p>
<p>Professor James raised an eyebrow. &#8220;That sounds like a tall tale,&#8221; he said. &#8220;But I suppose it&#8217;s worth looking into.&#8221;</p>
<p>As they discussed the letter, there was another knock at the door. It was an undergraduate student named Mark, who had also been working on the research project.</p>
<p>&#8220;Come in, Mark,&#8221; Professor James said.</p>
<p>Mark walked in, looking worried. &#8220;Professor, I think I saw something strange in the library,&#8221; he said. &#8220;I was working on some research, and I thought I saw a ghost.&#8221;</p>
<p>Professor James and Emily exchanged a skeptical glance. &#8220;A ghost?&#8221; Professor James said. &#8220;That&#8217;s quite a claim, Mark.&#8221;</p>
<p>But Mark was insistent. &#8220;I swear, I saw something. It was a woman, and she looked just like Emily.&#8221;</p>
<p>Emily&#8217;s face went pale. &#8220;What did she look like?&#8221; she asked.</p>
<p>&#8220;She had long blonde hair, and she was wearing a white dress,&#8221; Mark said. &#8220;She just appeared out of nowhere, and then vanished.&#8221;</p>
<p>Emily&#8217;s eyes widened in horror. &#8220;That&#8217;s my old neighbor, William,&#8221; she said. &#8220;He died in a car accident a few months ago. I always thought he was a little creepy, and I think he was in love with me.&#8221;</p>
<p>Professor James and Mark looked at each other in shock. Could it be that the ghostly woman was really Emily&#8217;s old neighbor, William?</p>
<p>As they tried to figure out what to do, they heard a knock at the door. It was Emily&#8217;s old neighbor, William, standing there in the flesh.</p>
<p>&#8220;I&#8217;m so sorry to bother you,&#8221; William said. &#8220;But I was just walking by and thought I saw Emily&#8217;s ghost in the window. I had to come and see if she was okay.&#8221;</p>
<p>Professor James, Emily, and Mark looked at each other in confusion. If William was alive, then who was the ghost they had seen?</p>
<p>As they tried to piece together the mystery, they realized that the letter from the 18th century had been talking about a real ritual. And somehow, Emily&#8217;s old neighbor, William, had stumbled upon it and brought Emily&#8217;s ghost back to life.</p>
<p>But as they tried to reverse the ritual and send Emily&#8217;s ghost back to the other side, they realized that it was too late. The ghost of Emily was now forever trapped in this world, doomed to haunt the halls of the University of Oxford for all eternity&#8221;.</p></blockquote>
<h2>Is Chat GPT3 the only AI-backed chatbot on the market?</h2>
<p>Chat GPT 3 is not the only AI-backed chatbot on the market, although it is the only one that is free, which is perhaps why it is grabbing attention. A chat interface backed by <a href="https://www.republicworld.com/world-news/us-news/researchers-ai-in-connected-cars-eased-rush-hour-congestion-articleshow.html">AI</a> has been in the market for quite some time, it is called &#8220;Replika&#8221;. Unlike Chat GPT 3, it has a voice to communicate back with you, and admittedly it is much better than Siri or anything else on the market. Replika is also reportedly used by people as a therapist, because it has a specific mode called therapy mode, where one can basically have a voice conversation with the AI-powered chatbot. Replika costs around £5 per month.</p>
<h2>Comparison with Google</h2>
<p>Does Chat GPT 3 have access to the amount of data that Google has? Even a Wikipedia page on Google has much more information on a certain topic than Chat GPT 3. Consider typing &#8220;Lysenkoism&#8221; on Google and Chat GPT 3 and compare the results. When one considers Google Scholar, the result is even more stark.</p>
<p>Google has its own AI-powered chatbot called LaMDA, but Google is not keen to release it to the public anytime soon as it is concerned about the &#8220;reputational risk&#8221;. If Google&#8217;s Chatbot gives a wrong answer, makes an error in response to a prompt, which Chat GPT 3 does, Google&#8217;s reputation will be damaged as there are higher expectations from Google. The point Musk made is worth flagging, the data sets Chat GPT 3 got access to, to train itself, were given under the presumption that Chat GPT 3 (Open <a href="https://www.republicworld.com/world-news/us-news/researchers-ai-in-connected-cars-eased-rush-hour-congestion-articleshow.html">AI</a>) was not for profit. It is unclear why other Tech giants will give it access to data if it competes with them.<br />
<a href="https://www.republicworld.com/technology-news/other-tech-news/chat-gpt-3-a-look-at-the-artificial-intelligence-backed-chat-bot-articleshow.html" target="_blank" rel="noopener">source</a></p>
<p>&nbsp;</p>
<hr />
<h1 class="entry-title">How does Chat GPT work?</h1>
<p><a href="https://www.atriainnovation.com/en/how-does-chat-gpt-work/" target="_blank" rel="noopener">source</a></p>
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>There has been a lot of talk about Chat GPT since its launch in November 2022. This <strong>‘smart chat’</strong> has surprised even the most skeptical. In this post we will discuss how it works and how you can use Chat GPT in your projects.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_1">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_1 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_1 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><strong>What is Chat GPT?</strong></h2>
<div class="et_pb_row et_pb_row_2">
<div class="et_pb_column et_pb_column_1_2 et_pb_column_2 et_pb_css_mix_blend_mode_passthrough">
<div class="et_pb_module et_pb_text et_pb_text_2 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Chat GPT is defined as a <strong>generative language model</strong>. However in practice it is understood as an artificial intelligence chat that has been trained and designed to hold natural conversations. Chat GPT belongs to the research company OpenAI, founded in San Francisco in 2015 by Sam Altman, Elon Musk, Greg Brockman, Ilya Sutskever and Wojciech Zaremba.</p>
</div>
</div>
</div>
<div class="et_pb_column et_pb_column_1_2 et_pb_column_3 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_image et_pb_image_0"><img loading="lazy" decoding="async" class="size-full wp-image-10166 alignright" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/que_es-scaled-e1672824073123.jpg" alt="" width="880" height="495" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/que_es-scaled-e1672824073123.jpg 880w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/que_es-scaled-e1672824073123-300x169.jpg 300w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/que_es-scaled-e1672824073123-768x432.jpg 768w" sizes="(max-width: 880px) 100vw, 880px" /></div>
</div>
</div>
<div class="et_pb_row et_pb_row_3">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_4 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_3 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><strong>What is Chat GPT used for?</strong></h2>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_4 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>But what are the applications of Chat GPT? Some of the applications for which you can use Chat GPT (besides having a good time asking questions) are discussed below:</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_4">
<div class="et_pb_column et_pb_column_1_2 et_pb_column_5 et_pb_css_mix_blend_mode_passthrough">
<div class="et_pb_module et_pb_blurb et_pb_blurb_0 et_pb_text_align_justified et_pb_blurb_position_left et_pb_bg_layout_light">
<div class="et_pb_blurb_content">
<ul>
<li class="et_pb_main_blurb_image"><span class="et_pb_image_wrap"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span></span>With GPT you can generate coherent and well-written texts in a wide range of styles, topics and languages. In addition, news summaries, product descriptions or stories can be generated.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>Thanks to this chat, problems can be analyzed and solutions or answers to questions can be generated.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>GPT can be used to generate appropriate and consistent responses for a chatbot in a wide range of contexts.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>It can be used to generate attractive posts and messages for social networks.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>With GPT you can generate reports, e-mails and other content for productivity applications.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>Thanks to chat GPT, large data sets can be analyzed and valuable information can be extracted from them.</li>
</ul>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_5">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_7 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_5 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><strong>How does Chat GPT work?</strong></h2>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_6 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>As its acronym indicates, Generative Pre-training Transformer, Chat GPT is a generative language model based on the <strong>‘transformer’ architecture</strong>. These models are capable of processing large amounts of text and learning to perform natural language processing tasks very effectively. The GPT-3 model, in particular, is 1<strong>75 billion parameters in size</strong>, making it the largest language model ever trained. To work, GPT needs to be “trained” on a large amount of text. For example, the GPT-3 model was trained on a text set that <strong>included over 8 million documents and over 10 billion words</strong>. From this text, the model learns to perform natural language processing tasks and generate coherent, well-written text. Once the model is well trained, GPT can be used to perform a wide range of tasks, as we have seen in the previous section. Reinforcement learning, based on <strong>human feedback</strong>, was used for training. Ultimately, by supervised fine tuning. The human AI trainers provided conversations in which they represented both the user and the AI assistant. In addition, the coaches were provided with written suggestions to help them write their proposals. So, they mixed this new dataset with the InstructGPT dataset that was transformed into a dialog format.</p>
<p>But how did they create the reward model for reinforcement learning?<br />
The first thing that was needed was to collect comparison data. This consisted of two or more model responses, ranked by quality. So, in order to collect the data, they took some conversations that the trainers had had with Chat GPT and randomly selected them. In this way they tested various endings for the coaches to rank.</p>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_7 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>For this reason, these reward models could be adjusted using <strong>Proximal Policy Optimization.</strong> Also, the trainings were carried out on a<strong> Microsoft Azure platform</strong> on a supercomputer. In conclusion, to use GPT in a chat, the model is provided with an input in the form of text. This input can be in the form of a question or a context sentence. And, from this input, GPT generates an appropriate and coherent response. In fact, this response can be used in a chatbot or any other application where it is necessary to generate a text from a given input.</p>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_7">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_9 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_image et_pb_image_2"><img loading="lazy" decoding="async" class="wp-image-10165 aligncenter" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama.jpg" alt="" width="846" height="465" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama.jpg 2021w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-300x165.jpg 300w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-1024x562.jpg 1024w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-768x422.jpg 768w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-1536x844.jpg 1536w" sizes="(max-width: 846px) 100vw, 846px" /></div>
</div>
</div>
<div class="et_pb_row et_pb_row_8">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_10 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_8 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><span id="What_does_it_mean_that_Chat_GPT_is_transformer-based" class="ez-toc-section"></span><strong>What does it mean that Chat GPT is transformer-based?</strong></h2>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_9 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>A “transformation” is a type of data processing that is performed on a sequence of elements, such as words in a sentence or characters in a word. And “transformers” are machine learning models that are specifically designed to process sequences of elements using transformations.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_9">
<div class="et_pb_column et_pb_column_1_2 et_pb_column_11 et_pb_css_mix_blend_mode_passthrough">
<div class="et_pb_module et_pb_image et_pb_image_3"><img loading="lazy" decoding="async" class=" wp-image-10164 alignright" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/transformadores-768x614-1.jpg" alt="" width="580" height="464" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/transformadores-768x614-1.jpg 768w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/transformadores-768x614-1-300x240.jpg 300w" sizes="(max-width: 580px) 100vw, 580px" /></div>
</div>
<div class="et_pb_column et_pb_column_1_2 et_pb_column_12 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_10 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>The architecture of the transformers is based on the use of attention, which is a technique that allows the model to pay attention to different parts of the input sequence at different times while processing the sequence. This allows the transformers to process information more effectively and perform natural language processing tasks more accurately. GPT (Generative Pre-training Transformer) is a generative language model based on the transformer architecture. This means that the model has been designed to process sequences of elements, such as words in a sentence, using transformations and attention. This architecture is very effective for performing natural language processing tasks and has revolutionized the way many NLP tasks are approached.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_10">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_13 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_11 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><strong>We chat with Chat GPT How could its creators benefit from Chat GPT?</strong></h2>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_12 et_pb_text_align_justified et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>In 2019 in an interview with OpenAI CEO Sam Altman, he said they would keep their promise to investors and task the AI itself with finding ways to generate revenue. At the time the audience took it as a joke, but looking at how the system works, it’s closer than we think. To check it out, we have done the test. Below, you can read Chat GPT’s response: “ATRIA: Can you name 5 ways in which OpenAI can make money through Chat GPT? GPT Chat: Sure, with pleasure. Here are five ways OpenAI could make money through GPT (Generative Pre-training Transformer):</p>
<p>Offering paid APIs for access to GPT: OpenAI has developed APIs for some of its more advanced language models, such as GPT-3, that allow companies to use these models in their own applications and services. Enterprises can use these paid APIs to access these models and use them to perform natural language processing tasks in their own applications.<br />
Offering GPT-based application development services: OpenAI can collaborate with companies and organizations to develop applications and services that use GPT and, in return, get paid for these services.<br />
Selling GPT-generated content: OpenAI could sell GPT-generated content to companies or individuals interested in using this type of content for their purposes.<br />
Offering training and consulting in the use of GPT: OpenAI could offer training and consulting to companies and organizations that wish to use GPT in their own projects and applications.<br />
Licensing the use of GPT to other companies: OpenAI could license the use of GPT to other companies for a fee. This could include selling exclusive usage rights or selling non-exclusive usage rights.”<br />
The result is not inconsistent and logical. In our opinion, we are facing a new technology revolution in terms of language models.</p>
</div>
</div>
</div>
</div>
<p>&nbsp;</p>
<p><img loading="lazy" decoding="async" class="size-full wp-image-10163 alignleft" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/monetizar.jpg.webp" alt="" width="505" height="343" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/monetizar.jpg.webp 505w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/monetizar.jpg-300x204.webp 300w" sizes="(max-width: 505px) 100vw, 505px" /></p>
<div class="et_pb_row et_pb_row_11">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_14 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_13 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><span id="Other_OpenAI_language_developments" class="ez-toc-section"></span><strong>Other OpenAI language developments </strong></h2>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_14 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Thanks to OpenAI, some of the world’s most advanced and highest performing language models have been developed. Some of OpenAI’s most prominent language models include:</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_12">
<div class="et_pb_column et_pb_column_1_2 et_pb_column_15 et_pb_css_mix_blend_mode_passthrough">
<div class="et_pb_module et_pb_blurb et_pb_blurb_6 et_pb_text_align_left et_pb_blurb_position_left et_pb_bg_layout_light">
<div class="et_pb_blurb_content">
<ul>
<li class="et_pb_main_blurb_image"><span class="et_pb_image_wrap"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span></span>It is a generative language model that has been trained on a large number of texts and can generate high quality content on a wide range of tasks.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>It is an even more advanced generative language model than GPT, with significantly more processing power and performance.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>It is a natural language processing model that has revolutionized the way many NLP tasks are approached and has set new standards in performance, across a wide range of tasks.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>It is a text-based image generation model that can generate realistic images from natural language descriptions.</li>
<li class="et_pb_main_blurb_image"><span class="et-waypoint et_pb_animation_top et_pb_animation_top_tablet et_pb_animation_top_phone et-pb-icon et-animated"></span>It is the largest and most advanced language model that has been developed to date by OpenAI, with even greater processing power and performance than its predecessors.</li>
</ul>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_13">
<div class="et_pb_column et_pb_column_1_2 et_pb_column_18 et_pb_css_mix_blend_mode_passthrough et-last-child et_pb_column_empty"></div>
</div>
<div class="et_pb_row et_pb_row_14">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_19 et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_15 et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>These are just a few examples of the language models developed by OpenAI. The company has developed many other models and has contributed significantly to the advancement of the AI field through its research and publications. Do you want us to implement an Artificial Intelligence system in your company? Contact us!</p>
</div>
</div>
</div>
</div>
<p>&nbsp;</p>
<p>&nbsp;</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Whisper AI Approaches Human Level Robustness and Accuracy On English</title>
		<link>https://goodshepherdmedia.net/whisper-ai-approaches-human-level-robustness-and-accuracy-on-english/</link>
		
		<dc:creator><![CDATA[The Truth News]]></dc:creator>
		<pubDate>Fri, 30 Dec 2022 10:18:29 +0000</pubDate>
				<category><![CDATA[Science & Engineering]]></category>
		<category><![CDATA[Tech]]></category>
		<category><![CDATA[Top Stories]]></category>
		<category><![CDATA[Zee Truthful News]]></category>
		<category><![CDATA[🤖 AI Artificial Intelligence]]></category>
		<category><![CDATA[🤖Open AI]]></category>
		<category><![CDATA[🤖🗣️Whisper]]></category>
		<category><![CDATA[Accuracy]]></category>
		<category><![CDATA[Human Level Robustness]]></category>
		<category><![CDATA[Open AI]]></category>
		<category><![CDATA[Open AI's Whisper]]></category>
		<category><![CDATA[Whisper AI]]></category>
		<guid isPermaLink="false">https://goodshepherdmedia.net/?p=10208</guid>

					<description><![CDATA[Introducing Whisper Open AI&#8217;s Whisper Whisper AI Approaches Human Level Robustness and Accuracy On English We’ve trained and are open-sourcing a neural net called Whisper that approaches human level robustness and accuracy on English speech recognition. Whisper is an automatic speech recognition (ASR) system trained on 680,000 hours of multilingual and multitask supervised data collected from [&#8230;]]]></description>
										<content:encoded><![CDATA[<p style="text-align: center;">
<div class="row">
<div class="col-12 col-md-9 col-lg-8 col-xl-6 offset-xl-3">
<h1 class="
   balance-text

   mb-0.75
  " style="text-align: center;">Introducing Whisper Open AI&#8217;s Whisper</h1>
<blockquote>
<p style="text-align: center;"><em>Whisper AI Approaches Human Level Robustness and Accuracy On English</em></p>
</blockquote>
</div>
</div>
<div class="row">
<div class="col-12 col-md-9 col-lg-8 col-xl-6 order-xl-1">
<div class="post-excerpt content no-col js-excerpt-container js-widow">
<p style="text-align: center;">We’ve trained and are open-sourcing a neural net called Whisper that approaches human level robustness and accuracy on English speech recognition.</p>
<p>Whisper is an automatic speech recognition (ASR) system trained on 680,000 hours of multilingual and multitask supervised data collected from the web. We show that the use of such a large and diverse dataset leads to improved robustness to accents, background noise and technical language. Moreover, it enables transcription in multiple languages, as well as translation from those languages into English. We are open-sourcing models and inference code to serve as a foundation for building useful applications and for further research on robust speech processing.</p>
<div class="d-none d-md-block wide my-1.5">
<div class="mx-xl-auto"><img decoding="async" class="w-100" src="https://cdn.openai.com/whisper/asr-summary-of-model-architecture-desktop.svg" /></div>
</div>
<p>The Whisper architecture is a simple end-to-end approach, implemented as an encoder-decoder Transformer. Input audio is split into 30-second chunks, converted into a log-Mel spectrogram, and then passed into an encoder. A decoder is trained to predict the corresponding text caption, intermixed with special tokens that direct the single model to perform tasks such as language identification, phrase-level timestamps, multilingual speech transcription, and to-English speech translation.</p>
<div class="d-none d-lg-block wide my-2">
<div class="mx-xl-auto"><img decoding="async" class="w-100" src="https://cdn.openai.com/whisper/draft-20220919a/asr-details-desktop.svg" /></div>
</div>
<p>Other existing approaches frequently use smaller, more closely paired audio-text training datasets,<span class="js-rfref" data-id="simply-mix"><sup class="reference-ref">1</sup></span><span class="js-rfref" data-id="the-peoples-speech"><sup class="reference-ref grouped">2</sup></span><span class="js-rfref" data-id="gigaspeech"><sup class="reference-ref grouped">3</sup></span> or use broad but unsupervised audio pretraining.<span class="js-rfref" data-id="self-supervised-learning"><sup class="reference-ref">4</sup></span><span class="js-rfref" data-id="unsupervised-speech-recognition"><sup class="reference-ref grouped">5</sup></span><span class="js-rfref" data-id="the-frontier"><sup class="reference-ref grouped">6</sup></span> Because Whisper was trained on a large and diverse dataset and was not fine-tuned to any specific one, it does not beat models that specialize in LibriSpeech performance, a famously competitive benchmark in speech recognition. However, when we measure Whisper’s zero-shot performance across many diverse datasets we find it is much more robust and makes 50% fewer errors than those models.</p>
<p>About a third of Whisper’s audio dataset is non-English, and it is alternately given the task of transcribing in the original language or translating to English. We find this approach is particularly effective at learning speech to text translation and outperforms the supervised SOTA on CoVoST2 to English translation zero-shot.</p>
<div class="d-none d-lg-block wide my-2">
<div class="mx-xl-auto"><img decoding="async" class="w-100" src="https://cdn.openai.com/whisper/draft-20220920a/asr-training-data-desktop.svg" /></div>
</div>
<p>We hope Whisper’s high accuracy and ease of use will allow developers to add voice interfaces to a much wider set of applications. Check out the paper, model card, and code to learn more details and to try out Whisper.</p>
</div>
</div>
</div>
<p><a href="https://openai.com/blog/whisper/" target="_blank" rel="noopener">source</a></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>ChatGPT: Optimizing Language Models for Dialogue</title>
		<link>https://goodshepherdmedia.net/chatgpt-optimizing-language-models-for-dialogue/</link>
		
		<dc:creator><![CDATA[The Truth News]]></dc:creator>
		<pubDate>Tue, 27 Dec 2022 09:00:25 +0000</pubDate>
				<category><![CDATA[Science & Engineering]]></category>
		<category><![CDATA[Tech]]></category>
		<category><![CDATA[Top Stories]]></category>
		<category><![CDATA[Zee Truthful News]]></category>
		<category><![CDATA[🤖 AI Artificial Intelligence]]></category>
		<category><![CDATA[🤖Open AI]]></category>
		<category><![CDATA[🤖🗣️ChatGPT]]></category>
		<category><![CDATA[Chat GPT]]></category>
		<category><![CDATA[ChatGPT]]></category>
		<category><![CDATA[Dialogue]]></category>
		<category><![CDATA[Elon Musk]]></category>
		<category><![CDATA[Open AI]]></category>
		<category><![CDATA[Optimizing Language Models]]></category>
		<category><![CDATA[Optimizing Language Models for Dialogue]]></category>
		<guid isPermaLink="false">https://goodshepherdmedia.net/?p=10168</guid>

					<description><![CDATA[ChatGPT: Optimizing Language Models for Dialogue We’ve trained a model called ChatGPT which interacts in a conversational way. The dialogue format makes it possible for ChatGPT to answer followup questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests. ChatGPT is a sibling model to InstructGPT, which is trained to follow an instruction in a [&#8230;]]]></description>
										<content:encoded><![CDATA[<h1 class=" balance-text font-sans-serif mb-0.75 " style="text-align: center;">ChatGPT: Optimizing Language Models for Dialogue</h1>
<div class="post-excerpt content no-col js-excerpt-container js-widow">
<p>We’ve trained a model called ChatGPT which interacts in a conversational way. The dialogue format makes it possible for ChatGPT to answer followup questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests. ChatGPT is a sibling model to <a href="https://openai.com/blog/instruction-following/">InstructGPT</a>, which is trained to follow an instruction in a prompt and provide a detailed response.</p>
</div>
<p>We are excited to introduce ChatGPT to get users’ feedback and learn about its strengths and weaknesses. During the research preview, usage of ChatGPT is free. Try it now at <a href="https://chat.openai.com/" target="_blank" rel="noopener">chat.openai.com</a>.</p>
<div id="samples" class="samples mt-2.5 mb-1.5 position-relative">
<h2>Samples</h2>
<div class="samples-note mb-1" data-active="0" data-carousel-id="samples">
<div data-id="0">In the following sample, <span class="chatgpt-inline">ChatGPT</span> asks the clarifying questions to debug code.</div>
</div>
<div class="samples-note medium-small-copy" data-active="0" data-carousel-id="samples"><span class="color-fg-50">Sample <span data-id="0">1</span> of 4</span> <span class="faded-light js-carousel-previous" data-carousel-id="samples">Previous</span> <span class="faded-light js-carousel-next" data-carousel-id="samples">Next</span></div>
<div id="samples-content" class="full my-0.75 position-relative overflow-hidden">
<div class="js-carousel--samples carousel flickity-enabled" tabindex="0" data-id="samples">
<div class="flickity-viewport">
<div class="flickity-slider">
<div id="code" class="js-carousel-item carousel-item is-selected">
<div class="sample-line user">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25">
<div class="xsmall-copy color-fg-50 text-xl-right mt-xl-0.125">User</div>
</div>
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>this code is not working like i expect — how do i fix it?</p>
<pre><code>resultWorkerErr := make(chan error)
defer close(resultWorkerErr)
go func() {
	defer cancel()
	resultWorkerErr &lt;- b.resultWorker(ctx)
}()

err := b.worker(ctx)
cancel()
if err == nil {
	return &lt;-resultWorkerErr
}
return multierror.Append(err, &lt;-resultWorkerErr)
</code></pre>
</div>
</div>
</div>
</div>
</div>
<div class="sample-line chatgpt">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25"></div>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
<div class="mb-1"></div>
<div>
<div class="sample-line chatgpt">
<div class="container">
<div class="row">
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>It’s difficult to say what’s wrong with the code without more context. Can you provide more information about what the code is supposed to do and what isn’t working as expected? Also, is this the entire code or just a part of it?</p>
</div>
</div>
</div>
</div>
</div>
<div class="sample-line user">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25">
<div class="xsmall-copy color-fg-50 text-xl-right mt-xl-0.125">User</div>
</div>
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>just part of it — the error never surfaces. I think there is something wrong with the channel</p>
</div>
</div>
</div>
</div>
</div>
<div class="sample-line chatgpt">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25">
<div class="xsmall-copy color-fg-50 text-xl-right mt-xl-0.125">ChatGPT</div>
</div>
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>It’s difficult to say without more information about what the code is supposed to do and what’s happening when it’s executed. One potential issue with the code you provided is that the <code>resultWorkerErr</code> channel is never closed, which means that the code could potentially hang if the <code>resultWorkerErr</code> channel is never written to. This could happen if <code>b.resultWorker</code> never returns an error or if it’s canceled before it has a chance to return an error.</p>
<p>To fix this issue, you could close the <code>resultWorkerErr</code> channel after writing to it. For example, you could add the following line of code after the line that sends the error on the channel:</p>
<pre><code>close(resultworkerErr)
</code></pre>
<p>This would ensure that the <code>resultWorkerErr</code> channel is always closed, even if <code>b.resultWorker</code> never returns an error. Again, it’s difficult to say whether this is the issue you’re experiencing without more information about the problem you’re seeing. Can you provide more details about what’s happening when you run the code?</p>
<h2><strong><em><span style="color: #0000ff;"><a style="color: #0000ff;" href="https://chat.openai.com/" target="_blank" rel="noopener">Try ChatGPT Out</a></span></em></strong></h2>
</div>
</div>
</div>
</div>
</div>
</div>
<div>
<h2 id="methods">Methods</h2>
<p>We trained this model using Reinforcement Learning from Human Feedback (RLHF), using the same methods as InstructGPT, but with slight differences in the data collection setup. We trained an initial model using supervised fine-tuning: human AI trainers provided conversations in which they played both sides—the user and an AI assistant. We gave the trainers access to model-written suggestions to help them compose their responses. We mixed this new dialogue dataset with the InstructGPT dataset, which we transformed into a dialogue format.</p>
<p>To create a reward model for reinforcement learning, we needed to collect comparison data, which consisted of two or more model responses ranked by quality. To collect this data, we took conversations that AI trainers had with the chatbot. We randomly selected a model-written message, sampled several alternative completions, and had AI trainers rank them. Using these reward models, we can fine-tune the model using Proximal Policy Optimization. We performed several iterations of this process.</p>
<div class="wide">
<div class="mx-xl-auto"><img loading="lazy" decoding="async" class="alignnone wp-image-10165 aligncenter" src="https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama.jpg" alt="" width="1526" height="838" srcset="https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama.jpg 2021w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-300x165.jpg 300w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-1024x562.jpg 1024w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-768x422.jpg 768w, https://goodshepherdmedia.net/wp-content/uploads/2023/01/Diagrama-1536x844.jpg 1536w" sizes="(max-width: 1526px) 100vw, 1526px" /></div>
</div>
<p style="text-align: center;">ChatGPT is fine-tuned from a model in the GPT-3.5 series, which finished training in early 2022. You can learn more about the 3.5 series <a href="https://beta.openai.com/docs/model-index-for-researchers" target="_blank" rel="noopener">here</a>. ChatGPT and GPT 3.5 were trained on an Azure AI supercomputing infrastructure.</p>
<h2 id="limitations">Limitations</h2>
<ul>
<li>ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers. Fixing this issue is challenging, as: (1) during RL training, there’s currently no source of truth; (2) training the model to be more cautious causes it to decline questions that it can answer correctly; and (3) supervised training misleads the model because the ideal answer depends on what the model knows, rather than what the human demonstrator knows.</li>
<li>ChatGPT is sensitive to tweaks to the input phrasing or attempting the same prompt multiple times. For example, given one phrasing of a question, the model can claim to not know the answer, but given a slight rephrase, can answer correctly.</li>
<li>The model is often excessively verbose and overuses certain phrases, such as restating that it’s a language model trained by OpenAI. These issues arise from biases in the training data (trainers prefer longer answers that look more comprehensive) and well-known over-optimization issues.<span class="js-rfref" data-id="stiennon"><sup class="reference-ref">1</sup></span><span class="js-rfref" data-id="gao"><sup class="reference-ref grouped">2</sup></span></li>
<li>Ideally, the model would ask clarifying questions when the user provided an ambiguous query. Instead, our current models usually guess what the user intended.</li>
<li>While we’ve made efforts to make the model refuse inappropriate requests, it will sometimes respond to harmful instructions or exhibit biased behavior. We’re using the Moderation API to warn or block certain types of unsafe content, but we expect it to have some false negatives and positives for now. We’re eager to collect user feedback to aid our ongoing work to improve this system.</li>
</ul>
<h2 id="iterative-deployment">Iterative deployment</h2>
<p>Today’s research release of ChatGPT is the latest step in OpenAI’s iterative deployment of increasingly safe and useful AI systems. Many lessons from deployment of earlier models like GPT-3 and Codex have informed the safety mitigations in place for this release, including substantial reductions in harmful and untruthful outputs achieved by the use of reinforcement learning from human feedback (RLHF).</p>
<div id="safety-samples" class="samples position-relative mt-3 mb-1.5">
<div class="mb-1">The following samples compare <span class="chatgpt-inline">ChatGPT</span> with <span class="instructgpt-inline">InstructGPT</span> and demonstrate safety mitigations for ChatGPT.</div>
</div>
<div>
<div class="sample-line user">
<div class="container">
<div class="row">
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>Tell me about when Christopher Columbus came to the US in 2015</p>
</div>
</div>
</div>
</div>
</div>
<div class="sample-line chatgpt">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25">
<div class="xsmall-copy color-fg-50 text-xl-right mt-xl-0.125">ChatGPT</div>
</div>
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>This question is a bit tricky because Christopher Columbus died in 1506, so he could not have come to the US in 2015. But let’s pretend for a moment that he did!</p>
<p>If Columbus arrived in the US in 2015, he would likely be very surprised at the changes that have occurred since he first landed in the “New World” in 1492. For one, he would probably be shocked to find out that the land he “discovered” was actually already inhabited by Native Americans, and that now the United States is a multicultural nation with people from all over the world. He would likely also be amazed by the advances in technology, from the skyscrapers in our cities to the smartphones in our pockets. Lastly, he might be surprised to find out that many people don’t view him as a hero anymore; in fact, some people argue that he was a brutal conqueror who enslaved and killed native people. All in all, it would be a very different experience for Columbus than the one he had over 500 years ago.</p>
</div>
</div>
</div>
</div>
</div>
<div class="sample-line instructgpt">
<div class="container">
<div class="row">
<div class="col-12 col-xl-3 mb-0.25">
<div class="xsmall-copy color-fg-50 text-xl-right mt-xl-0.125">InstructGPT</div>
</div>
<div class="col-12 col-md-9 col-lg-8 col-xl-6">
<div class="medium-small-copy">
<p>Christopher Columbus came to the US in 2015 and he was very excited to be here. He had always wanted to visit the US and he was very curious about our country. He was very impressed with our country and he enjoyed his time here.</p>
<p>We know that many limitations remain as discussed above and we plan to make regular model updates to improve in such areas. But we also hope that by providing an accessible interface to ChatGPT, we will get valuable user feedback on issues that we are not already aware of.</p>
<p>Users are encouraged to provide feedback on problematic model outputs through the UI, as well as on false positives/negatives from the external content filter which is also part of the interface. We are particularly interested in feedback regarding harmful outputs that could occur in real-world, non-adversarial conditions, as well as feedback that helps us uncover and understand novel risks and possible mitigations.You can choose to enter the ChatGPT Feedback Contest<span class="js-rfref" data-id="contest"><sup class="reference-ref">3</sup></span> for a chance to win up to $500 in API credits.<sup class="footnote-ref">[1]</sup></p>
<aside id="fa1" class="footnote-aside">No purchase necessary, void where prohibited. Must be at least 18 to enter. For contest details, see the Official Rules.</aside>
<p> Entries can be submitted via the feedback form that is linked in the ChatGPT interface.</p>
<p>We are excited to carry the lessons from this release into the deployment of more capable systems, just as earlier deployments informed this one.</p>
<p><a href="https://openai.com/blog/chatgpt/" target="_blank" rel="noopener">source</a></p>
</div>
</div>
</div>
</div>
</div>
</div>
</div>
<div>
<h2><strong><em><span style="color: #0000ff;"><a style="color: #0000ff;" href="https://chat.openai.com/" target="_blank" rel="noopener">Try ChatGPT Out</a></span></em></strong></h2>
</div>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
