Transcribing audio with One AI
Looking for a high-quality automatic audio transcription service? Try out One AI’s Smart Transcription, with powerful text-to-speech, Proofreading, Summary, and much more.
Whether your application involves customer support calls, interviews, meetings, brainstorming sessions, or podcasts, you can have them transcribed with high accuracy in just several minutes. And with a combination of Language Skills like proofreading or summarization, you could go even further! Now they can become customer feedback reports, automatic subtitles, doctor's prescriptions, and so on.
How can I try it?
Here is a nice example of how you can easily get all the necessary info from a video. Let's use this video from a TED talk. I used this website to download it. You can pick to download video or just audio, let’s move with the audio.
Now open One AI studio. On the bottom-right of the screen, you can find our Language Skills, represented by a bunch of rectangles. Each Skill can be dragged and dropped onto the Skills pipeline box, above the library.

Upload your audio file in .mp3 or .wav format. You’ll notice that the “Transcribing Audio” Language Skill is already picked. Now just press “Run the Pipeline” and after a few moments, you can see the results you can work with.

Let’s print the text and check what we got:
```
print(transcription.text)
```
How can I get it to work with my code?
At this point, you are just two steps away from inserting generated code into your project.
Step1: Pick one of the generated code you need to your code editor (click here to generate your own API key):

Step 2: Run “pip install oneai” for Python SDK or “npm install oneai” for Node.js SDK to get the library. Make sure to import all the required packages:
```
import oneai
import base64
oneai.api_key = "[YOUR ONEAI API KEY]"
```
That’s it, you’re good to go! Run your code and see the results.
```
with open("AudioFile.mp3", "rb") as f:
pipeline = oneai.Pipeline(
steps = [
oneai.skills.Transcribe()
]
)
transcription = pipeline.run(f)
```
Using Additional Language skills on recognized code.
Now, when we have transcription, we can use any other Skill. Let's say we want to get the topics discussed in the audio file. Just drag and drop the “Topics'' Skill and press “Run Pipeline”:
```
pipeline = oneai.Pipeline(
steps = [
oneai.skills.Topics(),
]
)
Topics = pipeline.run(transcription.text)
print(Topics.data[0].values)
```
The topics we got :
['Diversity', 'Complexity', 'Earth', 'Cloud', 'DNA', 'Energy', 'Universe']
What if we want to change the title of the video? Yeah, just drop the “Headlines” skill:
```
text = transcription.text
pipeline = oneai.Pipeline(
steps = [
oneai.skills.Headline()
]
)
Headline = pipeline.run(transcription.text)
print(Headline.data[0].values)
```
The new headline we got was: “The universe is not mush, but it's marsh“. Well, that’s an interesting thought 😊
As in our case with the historical context, we could extract all the famous names and dates to include in the report using only the following skills "Names" and "Numbers&Time". Of course, you don't need to run each skill individually. You can use all the skills together in a single pipeline like so:
```
with open("AudioFile.mp3", "rb") as f:
pipeline = oneai.Pipeline(
steps = [
oneai.skills.Transcribe(),
oneai.skills.Topics(),
oneai.skills.Headline(),
oneai.skills.Names(),
oneai.skills.Numbers(),
]
)
output = pipeline.run(f)
```
And the outputs will be:
Topics:
`print(Topics.data[0].values)`
`['Energy', 'Earth', 'Complexity', 'Cloud', 'Diversity', 'Universe', 'Dna']`
Headline:
`print(Headline.data[0].values)`
`["The universe is not mush, but it's marsh"]`
Names:
`[[x.name,x.value,x.timestamp] for x in output.transcription.names]`
Numbers and Time:
`[x.name,x.value] for x in output.transcription.numbers]`
Conclusion
No matter if you work for a large company or a small startup, you can now extract all the information you need from audio in a few minutes.
To get started, visit our Language Studio. Share your own ideas with the community.