Made in Vancouver, Canada by Picovoice
Leopard is an on-device speech-to-text engine. Leopard is:
- Private; All voice processing runs locally.
- Accurate
- Compact and Computationally-Efficient
- Cross-Platform:
- Linux (x86_64), macOS (x86_64, arm64), Windows (x86_64)
- Android and iOS
- Chrome, Safari, Firefox, and Edge
- Raspberry Pi (3, 4, 5)
AccessKey is your authentication and authorization token for deploying Picovoice SDKs, including Leopard. Anyone who is using Picovoice needs to have a valid AccessKey. You must keep your AccessKey secret. You would need internet connectivity to validate your AccessKey with Picovoice license servers even though the voice recognition is running 100% offline.
AccessKey also verifies that your usage is within the limits of your account. Everyone who signs up for
Picovoice Console receives the Free Tier
usage rights described
here. If you wish to increase your limits, you can purchase a subscription plan.
- English, French, German, Italian, Japanese, Korean, Portuguese, and Spanish.
- Support for additional languages is available for commercial customers on a case-by-case basis.
Install the demo package:
pip3 install pvleoparddemo
Run the following in the terminal:
leopard_demo_file --access_key ${ACCESS_KEY} --audio_paths ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you
wish to transcribe.
Build the demo:
cmake -S demo/c/ -B demo/c/build && cmake --build demo/c/build
Run the demo:
./demo/c/build/leopard_demo -a ${ACCESS_KEY} -l ${LIBRARY_PATH} -m ${MODEL_FILE_PATH} ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console, ${LIBRARY_PATH}
with the path to appropriate
library under lib, ${MODEL_FILE_PATH}
to path to default model file
(or your custom one), and ${AUDIO_FILE_PATH}
with a path to an audio file you wish to transcribe.
To run the demo, go to demo/ios/LeopardDemo and run:
pod install
Replace let accessKey = "${YOUR_ACCESS_KEY_HERE}"
in the file ViewModel.swift with your AccessKey
.
Then, using Xcode, open the generated LeopardDemo.xcworkspace
and run the application.
Using Android Studio, open demo/android/LeopardDemo as an Android project and then run the application.
Replace "${YOUR_ACCESS_KEY_HERE}"
in the file MainActivity.java with your AccessKey
.
Install the demo package:
yarn global add @picovoice/leopard-node-demo
Run the following in the terminal:
leopard-file-demo --access_key ${ACCESS_KEY} --input_audio_file_path ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you
wish to transcribe.
For more information about Node.js demos go to demo/nodejs.
To run the Leopard demo on Android or iOS with Flutter, you must have the Flutter SDK installed on your system. Once installed, you can run flutter doctor
to determine any other missing requirements for your relevant platform. Once your environment has been set up, launch a simulator or connect an Android/iOS device.
Run the prepare_demo
script from demo/flutter with a language code to set up the demo in the language of your choice (e.g. de
-> German, ko
-> Korean). To see a list of available languages, run prepare_demo
without a language code.
dart scripts/prepare_demo.dart ${LANGUAGE}
Replace "${YOUR_ACCESS_KEY_HERE}"
in the file main.dart with your AccessKey
.
Run the following command from demo/flutter to build and deploy the demo to your device:
flutter run
The demo requires cgo
, which on Windows may mean that you need to install a gcc compiler like Mingw to build it properly.
From demo/go run the following command from the terminal to build and run the file demo:
go run filedemo/leopard_file_demo.go -access_key "${ACCESS_KEY}" -input_audio_path "${AUDIO_FILE_PATH}"
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you wish to transcribe.
For more information about Go demos go to demo/go.
To run the React Native Leopard demo app you will first need to set up your React Native environment. For this, please refer to React Native's documentation. Once your environment has been set up, navigate to demo/react-native to run the following commands:
For Android:
yarn android-install # sets up environment
yarn android-run # builds and deploys to Android
For iOS:
yarn ios-install # sets up environment
yarn ios-run
The Leopard Java demo is a command-line application that lets you choose between running Leopard on an audio file or on microphone input.
From demo/java run the following commands from the terminal to build and run the file demo:
cd demo/java
./gradlew build
cd build/libs
java -jar leopard-file-demo.jar -a ${ACCESS_KEY} -i ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you wish to transcribe.
For more information about Java demos go to demo/java.
Leopard .NET demo is a command-line application that lets you choose between running Leopard on an audio file or on real-time microphone input.
From demo/dotnet/LeopardDemo run the following in the terminal:
dotnet run -c FileDemo.Release -- --access_key ${ACCESS_KEY} --input_audio_path ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you
wish to transcribe.
For more information about .NET demos, go to demo/dotnet.
Leopard Rust demo is a command-line application that lets you choose between running Leopard on an audio file or on real-time microphone input.
From demo/rust/filedemo run the following in the terminal:
cargo run --release -- --access_key ${ACCESS_KEY} --input_audio_path ${AUDIO_FILE_PATH}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
with a path to an audio file you
wish to transcribe.
For more information about Rust demos, go to demo/rust.
From demo/web run the following in the terminal:
yarn
yarn start
(or)
npm install
npm run start
Open http://localhost:5000
in your browser to try the demo.
From demo/react run the following in the terminal:
yarn
yarn start ${LANGUAGE}
(or)
npm install
npm run start ${LANGUAGE}
Open http://localhost:3000
in your browser to try the demo.
Install the Python SDK:
pip3 install pvleopard
Create an instance of the engine and transcribe an audio file:
import pvleopard
leopard = pvleopard.create(access_key='${ACCESS_KEY}')
print(leopard.process_file('${AUDIO_FILE_PATH}'))
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and
${AUDIO_FILE_PATH}
to path an audio file. Finally, when done be sure to explicitly release the resources using
leopard.delete()
.
Create an instance of the engine and transcribe an audio file:
#include <stdbool.h>
#include <stdio.h>
#include <stdlib.h>
#include "pv_leopard.h"
pv_leopard_t *leopard = NULL;
bool enable_automatic_punctuation = false;
bool enable_speaker_diarization = false;
pv_status_t status = pv_leopard_init(
"${ACCESS_KEY}",
"${MODEL_FILE_PATH}",
enable_automatic_punctuation,
enable_speaker_diarization,
&leopard);
if (status != PV_STATUS_SUCCESS) {
// error handling logic
}
char *transcript = NULL;
int32_t num_words = 0;
pv_word_t *words = NULL;
status = pv_leopard_process_file(
leopard,
"${AUDIO_FILE_PATH}",
&transcript,
&num_words,
&words);
if (status != PV_STATUS_SUCCESS) {
// error handling logic
}
fprintf(stdout, "%s\n", transcript);
for (int32_t i = 0; i < num_words; i++) {
fprintf(
stdout,
"[%s]\t.start_sec = %.1f .end_sec = %.1f .confidence = %.2f .speaker_tag = %d\n",
words[i].word,
words[i].start_sec,
words[i].end_sec,
words[i].confidence,
words[i].speaker_tag);
}
pv_leopard_transcript_delete(transcript);
pv_leopard_words_delete(words);
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console, ${MODEL_FILE_PATH}
to path to
default model file (or your custom one), and ${AUDIO_FILE_PATH}
to path an audio file.
Finally, when done be sure to release resources acquired using pv_leopard_delete(leopard)
.
The Leopard iOS binding is available via CocoaPods. To import it into your iOS project, add the following line to your Podfile and run pod install
:
pod 'Leopard-iOS'
Create an instance of the engine and transcribe an audio_file:
import Leopard
let modelPath = Bundle(for: type(of: self)).path(
forResource: "${MODEL_FILE}", // Name of the model file name for Leopard
ofType: "pv")!
let leopard = Leopard(accessKey: "${ACCESS_KEY}", modelPath: modelPath)
do {
let audioPath = Bundle(for: type(of: self)).path(forResource: "${AUDIO_FILE_NAME}", ofType: "${AUDIO_FILE_EXTENSION}")
let result = leopard.process(audioPath)
print(result.transcript)
} catch let error as LeopardError {
} catch { }
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console, ${MODEL_FILE}
a custom trained model from console or the default model, ${AUDIO_FILE_NAME}
with the name of the audio file and ${AUDIO_FILE_EXTENSION}
with the extension of the audio file.
To include the package in your Android project, ensure you have included mavenCentral()
in your top-level build.gradle
file and then add the following to your app's build.gradle
:
dependencies {
implementation 'ai.picovoice:leopard-android:${LATEST_VERSION}'
}
Create an instance of the engine and transcribe an audio file:
import ai.picovoice.leopard.*;
final String accessKey = "${ACCESS_KEY}"; // AccessKey obtained from Picovoice Console (https://console.picovoice.ai/)
final String modelPath = "${MODEL_FILE_PATH}";
try {
Leopard leopard = new Leopard.Builder()
.setAccessKey(accessKey)
.setModelPath(modelPath)
.build(appContext);
File audioFile = new File("${AUDIO_FILE_PATH}");
LeopardTranscript transcript = leopard.processFile(audioFile.getAbsolutePath());
} catch (LeopardException ex) { }
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console, ${MODEL_FILE_PATH}
with a custom trained model from console or the default model, and ${AUDIO_FILE_PATH}
with the path to the audio file.
Install the Node.js SDK:
yarn add @picovoice/leopard-node
Create instances of the Leopard class:
const Leopard = require("@picovoice/leopard-node");
const accessKey = "${ACCESS_KEY}" // Obtained from the Picovoice Console (https://console.picovoice.ai/)
let leopard = new Leopard(accessKey);
const result = engineInstance.processFile('${AUDIO_FILE_PATH}');
console.log(result.transcript);
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and
${AUDIO_FILE_PATH}
to path an audio file.
When done, be sure to release resources using release()
:
leopard.release();
Add the Leopard Flutter plugin to your pub.yaml.
dependencies:
leopard_flutter: ^<version>
Create an instance of the engine and transcribe an audio file:
import 'package:leopard/leopard.dart';
final String accessKey = '{ACCESS_KEY}' // AccessKey obtained from Picovoice Console (https://console.picovoice.ai/)
try {
Leopard _leopard = await Leopard.create(accessKey, '{MODEL_FILE_PATH}');
LeopardTranscript result = await _leopard.processFile("${AUDIO_FILE_PATH}");
print(result.transcript);
} on LeopardException catch (err) { }
Replace ${ACCESS_KEY}
with your AccessKey
obtained from Picovoice Console, ${MODEL_FILE_PATH}
with a custom trained model from Picovoice Console or the default model, and ${AUDIO_FILE_PATH}
with the path to the audio file.
Install the Go binding:
go get github.com/Picovoice/leopard/binding/go/v2
Create an instance of the engine and transcribe an audio file:
import . "github.com/Picovoice/leopard/binding/go/v2"
leopard = Leopard{AccessKey: "${ACCESS_KEY}"}
err := leopard.Init()
if err != nil {
// handle err init
}
defer leopard.Delete()
transcript, words, err := leopard.ProcessFile("${AUDIO_FILE_PATH}")
if err != nil {
// handle process error
}
log.Println(transcript)
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and
${AUDIO_FILE_PATH}
to path an audio file. Finally, when done be sure to explicitly release the resources using
leopard.Delete()
.
The Leopard React Native binding is available via NPM. Add it via the following command:
yarn add @picovoice/leopard-react-native
Create an instance of the engine and transcribe an audio file:
import {Leopard, LeopardErrors} from '@picovoice/leopard-react-native';
const getAudioFrame = () => {
// get audio frames
}
try {
const leopard = await Leopard.create("${ACCESS_KEY}", "${MODEL_FILE_PATH}")
const { transcript, words } = await leopard.processFile("${AUDIO_FILE_PATH}")
console.log(transcript)
} catch (err: any) {
if (err instanceof LeopardErrors) {
// handle error
}
}
Replace ${ACCESS_KEY}
with your AccessKey
obtained from Picovoice Console, ${MODEL_FILE_PATH}
with a custom trained model from Picovoice Console or the default model and ${AUDIO_FILE_PATH}
with the absolute path of the audio file.
When done be sure to explicitly release the resources using leopard.delete()
.
The latest Java bindings are available from the Maven Central Repository at:
ai.picovoice:leopard-java:${version}
Create an instance of the engine with the Leopard Builder class and transcribe an audio file:
import ai.picovoice.leopard.*;
final String accessKey = "${ACCESS_KEY}";
try {
Leopard leopard = new Leopard.Builder().setAccessKey(accessKey).build();
LeopardTranscript result = leopard.processFile("${AUDIO_FILE_PATH}");
leopard.delete();
} catch (LeopardException ex) { }
System.out.println(result.getTranscriptString());
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console and ${AUDIO_FILE_PATH}
to the path an audio file. Finally, when done be sure to explicitly release the resources using leopard.delete()
.
Install the .NET SDK using NuGet or the dotnet CLI:
dotnet add package Leopard
Create an instance of the engine and transcribe an audio file:
using Pv;
const string accessKey = "${ACCESS_KEY}";
const string audioPath = "${AUDIO_FILE_PATH}";
Leopard leopard = Leopard.Create(accessKey);
Console.Write(leopard.ProcessFile(audioPath));
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console. Finally, when done release the resources using leopard.Dispose()
.
First you will need Rust and Cargo installed on your system.
To add the leopard library into your app, add pv_leopard
to your app's Cargo.toml
manifest:
[dependencies]
pv_leopard = "*"
Create an instance of the engine using LeopardBuilder
instance and transcribe an audio file:
use leopard::LeopardBuilder;
fn main() {
let access_key = "${ACCESS_KEY}"; // AccessKey obtained from Picovoice Console (https://console.picovoice.ai/)
let leopard: Leopard = LeopardBuilder::new().access_key(access_key).init().expect("Unable to create Leopard");
if let Ok(leopard_transcript) = leopard.process_file("/absolute/path/to/audio_file") {
println!("{}", leopard_transcript.transcript);
}
}
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console.
Install the web SDK using yarn:
yarn add @picovoice/leopard-web
or using npm:
npm install --save @picovoice/leopard-web
Create an instance of the engine using LeopardWorker
and transcribe an audio file:
import { Leopard } from "@picovoice/leopard-web";
import leopardParams from "${PATH_TO_BASE64_LEOPARD_PARAMS}";
function getAudioData(): Int16Array {
// ... function to get audio data
return new Int16Array();
}
const leopard = await LeopardWorker.create(
"${ACCESS_KEY}",
{ base64: leopardParams },
);
const { transcript, words } = await leopard.process(getAudioData());
console.log(transcript);
console.log(words);
Replace ${ACCESS_KEY}
with yours obtained from Picovoice Console. Finally, when done release the resources using leopard.release()
.
yarn add @picovoice/leopard-react @picovoice/web-voice-processor
(or)
npm install @picovoice/leopard-react @picovoice/web-voice-processor
import { useLeopard } from "@picovoice/leopard-react";
function App(props) {
const {
result,
isLoaded,
error,
init,
processFile,
startRecording,
stopRecording,
isRecording,
recordingElapsedSec,
release,
} = useLeopard();
const initEngine = async () => {
await init(
"${ACCESS_KEY}",
leopardModel,
);
};
const handleFileUpload = async (audioFile: File) => {
await processFile(audioFile);
}
const toggleRecord = async () => {
if (isRecording) {
await stopRecording();
} else {
await startRecording();
}
};
useEffect(() => {
if (result !== null) {
console.log(result.transcript);
console.log(result.words);
}
}, [result])
}
- Added speaker diarization feature
- Added React SDK
- Improvements to error reporting
- Upgrades to authorization and authentication system
- Improved engine accuracy
- Various bug fixes and improvements
- Node min support bumped to Node 16
- Bumped iOS support to iOS 13+
- Patches to .NET support
- Added language support for French, German, Italian, Japanese, Korean, Portuguese and Spanish
- Added support for .NET 7.0 and fixed support for .NET Standard 2.0
- iOS minimum support moved to 11.0
- Improved stability and performance
- Added true-casing by default for transcription results
- Added option to enable automatic punctuation insertion
- Word timestamps and confidence returned as part of transcription
- Support for 3gp (AMR) and MP4/m4a (AAC) audio files
- Leopard Web SDK release
- Initial release