Skip to content

Commit b0594df

Browse files
committed
chore: pnpm format
1 parent af57432 commit b0594df

File tree

173 files changed

+13957
-15280
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

173 files changed

+13957
-15280
lines changed

README.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,4 +41,3 @@ Run the following command at the root of the folder to spin up a local dev serve
4141
```
4242
fern docs dev
4343
```
44-

asyncapi.json

Lines changed: 18 additions & 65 deletions
Original file line numberDiff line numberDiff line change
@@ -54,23 +54,17 @@
5454
},
5555
"query": {
5656
"type": "object",
57-
"required": [
58-
"sample_rate"
59-
],
57+
"required": ["sample_rate"],
6058
"properties": {
6159
"sample_rate": {
6260
"description": "The sample rate of the streamed audio",
6361
"type": "integer",
64-
"examples": [
65-
16000
66-
]
62+
"examples": [16000]
6763
},
6864
"word_boost": {
6965
"description": "Add up to 2500 characters of custom vocabulary.\nThe parameter value must be a JSON encoded array of strings. The JSON must be URL encoded like other query string parameters.\n",
7066
"type": "string",
71-
"examples": [
72-
"[\"foo\",\"bar\"]"
73-
]
67+
"examples": ["[\"foo\",\"bar\"]"]
7468
},
7569
"encoding": {
7670
"description": "The encoding of the audio data",
@@ -80,9 +74,7 @@
8074
"description": "Authenticate using a [generated temporary token](https://www.assemblyai.com/docs/speech-to-text/streaming#authenticate-with-a-temporary-token)\n",
8175
"type": "string",
8276
"format": "password",
83-
"examples": [
84-
"eyJhbGciOiJIUzI1..."
85-
]
77+
"examples": ["eyJhbGciOiJIUzI1..."]
8678
},
8779
"disable_partial_transcripts": {
8880
"summary": "Disable partial transcripts",
@@ -300,9 +292,7 @@
300292
"RealtimeBaseMessage": {
301293
"type": "object",
302294
"x-fern-sdk-group-name": "realtime",
303-
"required": [
304-
"message_type"
305-
],
295+
"required": ["message_type"],
306296
"properties": {
307297
"message_type": {
308298
"description": "Describes the type of the message",
@@ -338,9 +328,7 @@
338328
"type": "object",
339329
"x-fern-sdk-group-name": "realtime",
340330
"additionalProperties": false,
341-
"required": [
342-
"error"
343-
],
331+
"required": ["error"],
344332
"properties": {
345333
"error": {
346334
"type": "string"
@@ -366,10 +354,7 @@
366354
"RealtimeTranscriptType": {
367355
"type": "string",
368356
"x-fern-sdk-group-name": "realtime",
369-
"enum": [
370-
"PartialTranscript",
371-
"FinalTranscript"
372-
]
357+
"enum": ["PartialTranscript", "FinalTranscript"]
373358
},
374359
"RealtimeTranscript": {
375360
"x-fern-sdk-group-name": "realtime",
@@ -392,11 +377,7 @@
392377
},
393378
{
394379
"type": "object",
395-
"required": [
396-
"message_type",
397-
"session_id",
398-
"expires_at"
399-
],
380+
"required": ["message_type", "session_id", "expires_at"],
400381
"properties": {
401382
"message_type": {
402383
"description": "Describes the type of the message",
@@ -437,10 +418,7 @@
437418
},
438419
{
439420
"type": "object",
440-
"required": [
441-
"message_type",
442-
"audio_duration_seconds"
443-
],
421+
"required": ["message_type", "audio_duration_seconds"],
444422
"properties": {
445423
"message_type": {
446424
"description": "Describes the type of the message",
@@ -472,9 +450,7 @@
472450
},
473451
{
474452
"type": "object",
475-
"required": [
476-
"message_type"
477-
],
453+
"required": ["message_type"],
478454
"properties": {
479455
"message_type": {
480456
"description": "Describes the type of the message",
@@ -548,9 +524,7 @@
548524
},
549525
{
550526
"type": "object",
551-
"required": [
552-
"message_type"
553-
],
527+
"required": ["message_type"],
554528
"properties": {
555529
"message_type": {
556530
"description": "Describes the type of message",
@@ -607,11 +581,7 @@
607581
},
608582
{
609583
"type": "object",
610-
"required": [
611-
"message_type",
612-
"punctuated",
613-
"text_formatted"
614-
],
584+
"required": ["message_type", "punctuated", "text_formatted"],
615585
"properties": {
616586
"message_type": {
617587
"description": "Describes the type of message",
@@ -714,12 +684,7 @@
714684
"Word": {
715685
"type": "object",
716686
"x-fern-sdk-group-name": "realtime",
717-
"required": [
718-
"start",
719-
"end",
720-
"confidence",
721-
"text"
722-
],
687+
"required": ["start", "end", "confidence", "text"],
723688
"properties": {
724689
"start": {
725690
"description": "Start time of the word in milliseconds",
@@ -798,9 +763,7 @@
798763
"description": "Manually end an utterance",
799764
"x-fern-sdk-group-name": "realtime",
800765
"type": "object",
801-
"required": [
802-
"force_end_utterance"
803-
],
766+
"required": ["force_end_utterance"],
804767
"properties": {
805768
"force_end_utterance": {
806769
"description": "A boolean value to communicate that you wish to force the end of the utterance",
@@ -818,9 +781,7 @@
818781
"description": "Configure the threshold for how long to wait before ending an utterance. Default is 700ms.",
819782
"x-fern-sdk-group-name": "realtime",
820783
"type": "object",
821-
"required": [
822-
"end_utterance_silence_threshold"
823-
],
784+
"required": ["end_utterance_silence_threshold"],
824785
"properties": {
825786
"end_utterance_silence_threshold": {
826787
"description": "The duration threshold in milliseconds",
@@ -840,9 +801,7 @@
840801
"summary": "Terminate session",
841802
"x-fern-sdk-group-name": "realtime",
842803
"type": "object",
843-
"required": [
844-
"terminate_session"
845-
],
804+
"required": ["terminate_session"],
846805
"properties": {
847806
"terminate_session": {
848807
"description": "Set to true to end your streaming session forever",
@@ -861,10 +820,7 @@
861820
"x-fern-sdk-group-name": "realtime",
862821
"description": "The encoding of the audio data",
863822
"default": "pcm_s16le",
864-
"enum": [
865-
"pcm_s16le",
866-
"pcm_mulaw"
867-
],
823+
"enum": ["pcm_s16le", "pcm_mulaw"],
868824
"x-fern-enum": {
869825
"pcm_s16le": {
870826
"description": "PCM signed 16-bit little-endian",
@@ -885,10 +841,7 @@
885841
}
886842
}
887843
},
888-
"examples": [
889-
"pcm_s16le",
890-
"pcm_mulaw"
891-
]
844+
"examples": ["pcm_s16le", "pcm_mulaw"]
892845
}
893846
},
894847
"securitySchemes": {
Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,14 @@
11
# Audio Intelligence 🤖
2+
23
Use our Audio Intelligence models to analyze audio and gain additional insights beyond speech to text.
34

45
## All Audio Intelligence Cookbooks
56

6-
| Model | Cookbooks |
7-
|----------------|-----------------------------------|
8-
| **Content Moderation** | [Identify Hate Speech in Audio and Video Files](content_moderation.ipynb) |
9-
| **Entity Detection** | [Create a Redacted Transcript with Entity Detection](entity_redaction.ipynb) |
10-
| **Auto Chapters** | [Create Summarized Chapters from Podcasts](auto_chapters.ipynb) |
11-
| **Summarization** | [Summarize Virtual Meetings](summarization.ipynb) |
12-
| **Topic Detection** | [Label Content with Topic Tags](topic_detection.ipynb) |
13-
| **Key Phrases** | [Identify Highlights in Audio and Video Files](key_phrases.ipynb) |
7+
| Model | Cookbooks |
8+
| ---------------------- | ---------------------------------------------------------------------------- |
9+
| **Content Moderation** | [Identify Hate Speech in Audio and Video Files](content_moderation.ipynb) |
10+
| **Entity Detection** | [Create a Redacted Transcript with Entity Detection](entity_redaction.ipynb) |
11+
| **Auto Chapters** | [Create Summarized Chapters from Podcasts](auto_chapters.ipynb) |
12+
| **Summarization** | [Summarize Virtual Meetings](summarization.ipynb) |
13+
| **Topic Detection** | [Label Content with Topic Tags](topic_detection.ipynb) |
14+
| **Key Phrases** | [Identify Highlights in Audio and Video Files](key_phrases.ipynb) |

cookbook-master/core-transcription/README.md

Lines changed: 35 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
11
# Core Transcription 🎙️
2+
23
The Speech Recognition model enables you to transcribe spoken words into written text and is the foundation of all AssemblyAI products.
34
On top of the core transcription, you can enable other features and models, such as [Speaker Diarization](https://www.assemblyai.com/docs/speech-to-text/speaker-diarization), by adding additional parameters to the same transcription request.
45

@@ -20,31 +21,38 @@ On top of the core transcription, you can enable other features and models, such
2021
- [Migration Guides](#migration-guides)
2122
- [Do More with our SDKS](#do-more-with-our-sdks)
2223

23-
2424
## All Core Transcription Cookbooks
2525

2626
<a name="basic"></a>
27+
2728
### Basic Transcription Workflows
28-
[Transcribe an Audio File](transcribe.ipynb)<br>
29+
30+
[Transcribe an Audio File](transcribe.ipynb)<br>
2931
[Specify a Language](specify-language.ipynb)
3032
[Transcribe YouTube videos](transcribe_youtube_videos.ipynb)<br>
3133
[Build a UI for Transcription with Gradio](gradio-frontend.ipynb)
32-
[Detect Low Confidence Words in a Transcript](detecting-low-confidence-words.md)
34+
[Detect Low Confidence Words in a Transcript](detecting-low-confidence-words.md)
3335

3436
<a name="batch"></a>
37+
3538
### Batch Transcription
36-
[Transcribe a batch of files using AssemblyAI](transcribe_batch_of_files)
37-
[Transcribe multiple files simultaneously using our Python SDK](SDK_transcribe_batch_of_files/batch_transcription.ipynb)
38-
[Transcribe multiple files simultaneously using our Node.js SDK](SDK-Node-batch.md)
39+
40+
[Transcribe a batch of files using AssemblyAI](transcribe_batch_of_files)
41+
[Transcribe multiple files simultaneously using our Python SDK](SDK_transcribe_batch_of_files/batch_transcription.ipynb)
42+
[Transcribe multiple files simultaneously using our Node.js SDK](SDK-Node-batch.md)
3943

4044
<a name="host-files"></a>
45+
4146
### Hosting Audio Files
47+
4248
[Transcribe from an AWS S3 Bucket](transcribe_from_s3.ipynb)
4349
[Transcribe Google Drive links](transcribing-google-drive-file.md)<br>
44-
[Transcribe GitHub Files](transcribing-github-files.md)
50+
[Transcribe GitHub Files](transcribing-github-files.md)
4551

4652
<a name="speaker-labels"></a>
53+
4754
### Speaker Labels
55+
4856
[Identify Speakers in Audio Recordings](speaker_labels.ipynb)<br>
4957
[Generate Speaker Labels with Make.com](make.com-speaker-labels.md)\
5058
[Calculate Talk/Listen Ratio of Speakers](talk-listen-ratio.ipynb)<br>
@@ -54,26 +62,34 @@ On top of the core transcription, you can enable other features and models, such
5462
[Speaker Identification Across Files w/ AssemblyAI, Pinecone, and Nvidia's TitaNet Model](titanet-speaker-identification.ipynb)
5563

5664
<a name="ald"></a>
65+
5766
### Automatic Language Detection
58-
[Use Automatic Language Detection](automatic-language-detection.ipynb)
59-
[Automatic Language Detection as separate step from Transcription](automatic-language-detection-separate.ipynb)
67+
68+
[Use Automatic Language Detection](automatic-language-detection.ipynb)
69+
[Automatic Language Detection as separate step from Transcription](automatic-language-detection-separate.ipynb)
6070
[Route to Default Language if Language Detection Confidence is Low - JS](automatic-language-detection-route-default-language-js.md)\
6171
[Route to Default Language if Language Detection Confidence is Low - Python](automatic-language-detection-route-default-language-python.ipynb)<br>
6272
[Route to Nano Speech Model if Language Confidence is Low](automatic-language-detection-route-nano-model.ipynb)
6373

6474
<a name="subtitles"></a>
75+
6576
### Subtitles
77+
6678
[Generate Subtitles for Videos](subtitles.ipynb)\
6779
[Create Subtitles with Speaker Labels](speaker_labelled_subtitles.ipynb)<br>
6880
[Create custom-length subtitles with AssemblyAI](subtitle_creation_by_word_count.ipynb)
6981

7082
<a name="delete"></a>
83+
7184
### Delete Transcripts
85+
7286
[Delete a Transcript ](delete_transcript.ipynb)
73-
[Delete transcripts after 24 hours of creation](schedule_delete.ipynb)
87+
[Delete transcripts after 24 hours of creation](schedule_delete.ipynb)
7488

7589
<a name="errors"></a>
90+
7691
### Error Handling and Audio File Fixes
92+
7793
[Troubleshoot common errors when starting to use our API](common_errors_and_solutions.md)<br>
7894
[Automatically Retry Server Errors](retry-server-error.ipynb)
7995
[Automatically Retry Upload Errors](retry-upload-error.ipynb)\
@@ -82,25 +98,32 @@ On top of the core transcription, you can enable other features and models, such
8298
](audio-duration-fix.ipynb)
8399

84100
<a name="translate"></a>
101+
85102
### Translation
103+
86104
[Translate Transcripts](translate_transcripts.ipynb)
87105
[Translate Subtitles](translate_subtitles.ipynb)
88106

89107
<a name="chunking"></a>
108+
90109
### Async Chunking for Near-Realtime Transcription
110+
91111
🆕 [Near-Realtime Python Speech-to-Text App](https://github.com/AssemblyAI-Solutions/async-chunk-py)\
92112
🆕 [Near-Realtime Node.js Speech-to-Text App](https://github.com/AssemblyAI-Solutions/async-chunk-js)\
93113
[Split audio file to shorter files](split_audio_file)
94114

95115
<a name="migration-guides"></a>
116+
96117
### Migration Guides
118+
97119
🆕 [AWS Transcribe to AssemblyAI](migration_guides/aws_to_aai.ipynb)\
98-
🆕 [Deepgram to AssemblyAI](migration_guides/dg_to_aai.ipynb)<br>
120+
🆕 [Deepgram to AssemblyAI](migration_guides/dg_to_aai.ipynb)<br>
99121
🆕 [OpenAI to AssemblyAI](migration_guides/oai_to_aai.ipynb)<br>
100122
🆕 [Google to AssemblyAI](migration_guides/google_to_aai.ipynb)
101123

102-
103124
<a name="do-more-with-sdk"></a>
125+
104126
### Do More with our SDKS
127+
105128
[Do more with the JavaScript SDK](do-more-with-sdk-js.md)\
106129
[Do more with the Python SDK](do-more-with-sdk-python.ipynb)

0 commit comments

Comments
 (0)