LeMUR

Summarize your audio data

1import com.assemblyai.api.AssemblyAI;
2import com.assemblyai.api.resources.lemur.requests.LemurTaskParams;
3import com.assemblyai.api.resources.transcripts.types.Transcript;
4import java.util.List;
5
6public final class App {
7 public static void main(String... args) {
8
9 AssemblyAI client = AssemblyAI.builder()
10 .apiKey("<YOUR_API_KEY>")
11 .build();
12
13 // You can use a local file:
14 /*
15 Transcript transcript = aai.transcripts().transcribe(
16 new File("./example.mp3"), params);
17 */
18
19 // Or use a publicly-accessible URL:
20 String audioUrl = "https://assembly.ai/sports_injuries.mp3";
21 Transcript transcript = client.transcripts().transcribe(audioUrl);
22
23 var params = LemurTaskParams.builder()
24 .prompt("Provide a brief summary of the transcript.")
25 .transcriptIds(List.of(transcript.getId()))
26 .finalModel(LemurModel.ANTHROPIC_CLAUDE3_5_SONNET)
27 .build();
28
29 var result = client.lemur().task(params);
30
31 System.out.println(result.getResponse());
32 }
33}

If you run the code above, you’ll see the following output:

1The transcript describes several common sports injuries - runner's knee,
2sprained ankle, meniscus tear, rotator cuff tear, and ACL tear. It provides
3definitions, causes, and symptoms for each injury. The transcript seems to be
4narrating sports footage and describing injuries as they occur to the athletes.
5Overall, it provides an overview of these common sports injuries that can result
6from overuse or sudden trauma during athletic activities

Ask questions about your audio data

Q&A with the task endpoint

To ask question about your audio data, define a prompt with your questions and call client.lemur().task(). Use the transcriptIds parameter to send one or more transcripts as additional context for the model.

1import com.assemblyai.api.AssemblyAI;
2import com.assemblyai.api.resources.transcripts.types.*;
3import com.assemblyai.api.resources.lemur.requests.*;
4import java.util.List;
5
6public final class App {
7 public static void main(String[] args) {
8
9 AssemblyAI client = AssemblyAI.builder()
10 .apiKey("<YOUR_API_KEY>")
11 .build();
12
13 // Step 1: Transcribe an audio file. For local files see our Getting Started guides.
14 String audioUrl = "https://assembly.ai/sports_injuries.mp3";
15 Transcript transcript = client.transcripts().transcribe(audioUrl);
16
17 // Step 2: Define a prompt with your question(s).
18 String prompt = "What is a runner's knee?";
19
20 // Step 3: Apply LeMUR.
21 var params = LemurTaskParams.builder()
22 .prompt(prompt)
23 .transcriptIds(List.of(transcript.getId()))
24 .finalModel(LemurModel.ANTHROPIC_CLAUDE3_5_SONNET)
25 .build();
26
27 var response = client.lemur().task(params);
28
29 System.out.println(response.getResponse());
30 }
31}

#Example output

1Based on the transcript, runner's knee is a condition characterized
2by pain behind or around the kneecap. It is caused by overuse,
3muscle imbalance and inadequate stretching. Symptoms include pain
4under or around the kneecap and pain when walking.

Q&A with the question-answer endpoint

The LeMUR Question & Answer function requires no prompt engineering and facilitates more deterministic and structured outputs. See the code examples below for more information on how to use this endpoint.

To use it, define a list of LemurQuestion objects. For each question, you can define additional context and specify either a answerFormat or a list of answerOptions. Additionally, you can define an overall context.

1import com.assemblyai.api.AssemblyAI;
2import com.assemblyai.api.resources.transcripts.types.*;
3import com.assemblyai.api.resources.lemur.requests.*;
4import java.util.List;
5
6public final class App {
7 public static void main(String[] args) {
8
9 AssemblyAI client = AssemblyAI.builder()
10 .apiKey("<YOUR_API_KEY>")
11 .build();
12
13 String audioUrl = "https://assembly.ai/meeting.mp4";
14
15 Transcript transcript = client.transcripts().transcribe(audioUrl);
16
17 var question1 = LemurQuestion.builder()
18 .question("What are the top level KPIs for engineering?")
19 .context(LemurQuestionContext.of("KPI stands for key performance indicator"))
20 .answerFormat("short sentence").build();
21
22 var question2 = LemurQuestion.builder()
23 .question("How many days has it been since the data team has gotten updated metrics?")
24 .answerOptions(List.of("1", "2", "3", "4", "5", "6", "7", "more than 7")).build();
25
26 var response = client.lemur().questionAnswer(LemurQuestionAnswerParams.builder()
27 .transcriptIds(List.of(transcript.getId()))
28 .finalModel(LemurModel.ANTHROPIC_CLAUDE3_5_SONNET)
29 .context(LemurBaseParamsContext.of("A GitLab meeting to discuss logistic"))
30 .questions(List.of(question1, question2))
31 .build());
32
33 for (var qa : response.getResponse()) {
34 System.out.println("Question: " + qa.getQuestion());
35 System.out.println("Answer: " + qa.getAnswer());
36 }
37 }
38}

For the full API reference, as well as the supported models and FAQs, refer to the full LeMUR Q&A guide.

Change the model type

LeMUR features the following LLMs:

  • Claude 3.5 Sonnet
  • Claude 3 Opus
  • Claude 3 Haiku
  • Claude 3 Sonnet

You can switch the model by specifying the final_model parameter.

1var params = LemurTaskParams.builder()
2 .prompt(prompt)
3 .transcriptIds(List.of(transcript.getId()))
4 .finalModel(LemurModel.ANTHROPIC_CLAUDE3_5_SONNET)
5 .build();
ModelSDK ParameterDescription
Claude 3.5 SonnetLemurModel.ANTHROPIC_CLAUDE3_5_SONNETClaude 3.5 Sonnet is the most intelligent model to date, outperforming Claude 3 Opus on a wide range of evaluations, with the speed and cost of Claude 3 Sonnet. This uses Anthropic’s Claude 3.5 Sonnet model version claude-3-5-sonnet-20240620.
Claude 3.0 OpusLemurModel.ANTHROPIC_CLAUDE3_OPUSClaude 3 Opus is good at handling complex analysis, longer tasks with many steps, and higher-order math and coding tasks.
Claude 3.0 HaikuLemurModel.ANTHROPIC_CLAUDE3_HAIKUClaude 3 Haiku is the fastest model that can execute lightweight actions.
Claude 3.0 SonnetLemurModel.ANTHROPIC_CLAUDE3_SONNETClaude 3 Sonnet is a legacy model with a balanced combination of performance and speed for efficient, high-throughput tasks.

You can find more information on pricing for each model here.

Change the maximum output size

You can change the maximum output size in tokens by specifying the maxOutputSize parameter. Up to 4000 tokens are allowed.

1var params = LemurTaskParams.builder()
2 .prompt(prompt)
3 .transcriptIds(List.of(transcript.getId()))
4 .maxOutputSize(1000)
5 .build();

Change the temperature

You can change the temperature by specifying the temperature parameter, ranging from 0.0 to 1.0.

Higher values result in answers that are more creative, lower values are more conservative.

1var params = LemurTaskParams.builder()
2 .prompt(prompt)
3 .transcriptIds(List.of(transcript.getId()))
4 .temperature(0.7)
5 .build();

Send customized input

You can submit custom text inputs to LeMUR without transcript IDs. This allows you to customize the input, for example, you could include the speaker labels for the LLM.

To submit custom text input, use the .inputText() method instead of .transcriptIds().

1var params = TranscriptOptionalParams.builder()
2 .speakerLabels(true)
3 .build();
4
5Transcript transcript = client.transcripts().transcribe(audioUrl, params);
6
7String textWithSpeakerLabels = transcript.getUtterances()
8 .map(utterances -> utterances.stream()
9 .map(utterance -> "Speaker " + utterance.getSpeaker() + ":\n" + utterance.getText() + "\n")
10 .collect(Collectors.joining()))
11 .orElse("");
12
13var response = client.lemur().task(LemurTaskParams.builder()
14 .prompt(prompt)
15 .inputText(textWithSpeakerLabels)
16 .build());

Submit multiple transcripts

LeMUR can easily ingest multiple transcripts in a single API call.

You can feed in up to a maximum of 100 files or 100 hours, whichever is lower.

1var response = client.lemur().task(LemurTaskParams.builder()
2 .prompt(prompt)
3 .transcriptIds(List.of(id1, id2, id3))
4 .build());

Delete LeMUR request data

You can delete the data for a previously submitted LeMUR request.

Response data from the LLM, as well as any context provided in the original request will be removed.

1var response = client.lemur().task(LemurTaskParams.builder()
2 .prompt(prompt)
3 .transcriptIds(List.of(transcript.getId()))
4 .build());
5
6var deletionResponse = client.lemur().purgeRequestData(response.getRequestId());