Skip to main content

Documentation Summary

Complete SpeechLytics API Documentation

This documentation has been fully generated based on the Swagger API definition. Below is a comprehensive overview of all documented sections.

📋 Main Pages

1. Introduction (/docs/intro)

  • Overview of SpeechLytics platform
  • Key features and benefits
  • Quick start guide
  • Service tiers and capabilities

2. Authentication (/docs/authentication)

  • How to get authentication tokens
  • Token management and expiration
  • Code examples (Python, JavaScript, C#, cURL)
  • Security best practices
  • Error handling

🎤 Speech-to-Text API

Getting Started (/docs/speech-to-text/getting-started)

  • Step-by-step workflow
  • File preparation (Base64 encoding)
  • Uploading and transcribing audio
  • Status checking and polling
  • Retrieving results
  • Complete code examples

Features (/docs/speech-to-text/features)

  • Quality metrics and scoring
  • Audio type detection (mono/stereo)
  • Channel-specific analysis
  • Transcription formats
  • Speaker frequency analysis
  • Word frequency analysis
  • Call duration tracking
  • Keyword matching
  • Priority processing
  • Metadata support
  • Supported audio formats and languages

🔴 Live Transcription API

Getting Started (/docs/live-speech/getting-started)

  • Real-time transcription workflow
  • Session management
  • WebSocket and polling methods
  • Audio streaming
  • Payload retrieval
  • Session lifecycle
  • Python and JavaScript examples

Features (/docs/live-speech/features)

  • Multi-channel support
  • Network information tracking
  • Real-time payload streaming
  • Language support
  • Audio quality monitoring
  • Session metadata
  • Connection management
  • Error handling
  • Performance optimization
  • Security features
  • Scalability considerations

🧠 Audio Intelligence APIs

Sentiment Analysis (/docs/audio-inteligence/sentiment)

  • Sentiment types (positive, negative, neutral, mixed)
  • Getting sentiment data
  • Sentiment queries and filtering
  • Available sentiments endpoint
  • Sentiment confidence scoring
  • Speaker-specific sentiment
  • Code examples
  • Visualization
  • Use cases

Topic Detection (/docs/audio-inteligence/topics)

  • Topic identification and classification
  • Available topics endpoint
  • Topic-based filtering
  • Multi-topic conversations
  • Topic trend analysis
  • Custom topic categories
  • Topic routing
  • Integration examples
  • Visualization

Named Entity Recognition (/docs/audio-inteligence/named-entity-recognition)

  • Supported entity types (person, organization, location, date, time, phone, email, money, etc.)
  • Entity extraction from transcripts
  • Confidence scoring
  • Entity linking
  • Contact information extraction
  • Financial data extraction
  • Appointment scheduling
  • PII handling and compliance
  • Entity statistics and reporting

Call Summarization (/docs/audio-inteligence/summarization)

  • Automatic summary generation
  • Summary structure and content
  • Getting summaries from transcripts
  • Action item extraction
  • Resolution tracking
  • Quality factors
  • CRM integration
  • Use cases and benefits

Translation (/docs/audio-inteligence/translation)

  • Multi-language support (100+ languages)
  • Transcript translation
  • Translation confidence scoring
  • Domain-specific translation
  • Professional review workflows
  • Multilingual QA processes
  • Use cases
  • Compliance considerations

📊 Analytics & Insights

Insights API (/docs/insights)

  • Available topics endpoint
  • Available sentiments endpoint
  • Call insights and analytics
  • Topic and sentiment distribution
  • Quality trends
  • Daily, weekly, and monthly reporting
  • Dashboard integration
  • Analytics client library
  • Advanced analysis capabilities

🏥 Health Monitoring

Health Check API (/docs/health-check)

  • System health endpoint
  • Service component status
  • Queue management
  • Processing time metrics
  • Error codes and troubleshooting
  • Monitoring implementation
  • Alert configuration
  • Capacity planning
  • Dashboard integration

📚 Documentation Structure

By Use Case

Getting Started with Transcription:

  1. Authentication
  2. Speech-to-Text Getting Started
  3. Speech-to-Text Features

Real-Time Transcription:

  1. Live Transcription Getting Started
  2. Live Transcription Features

Extracting Insights:

  1. Sentiment Analysis
  2. Topic Detection
  3. Named Entity Recognition

Advanced Processing:

  1. Call Summarization
  2. Translation

Analytics & Monitoring:

  1. Insights
  2. Health Check

🔗 API Endpoints Documented

Authentication

  • POST /api/v1/auth/token - Create authentication token

Speech-to-Text

  • POST /api/v1/transcribe - Upload and transcribe audio
  • GET /api/v1/transcripts/{id}/status - Get transcript status
  • GET /api/v1/transcripts - List transcripts with filtering

Live Transcription

  • POST /api/v1/live-transcribe/start - Start live session
  • GET /api/v1/live-transcribe/{id}/payload - Get live payload
  • POST /api/v1/live-transcribe/stop - Stop live session

Analytics

  • GET /api/v1/insights/topics - Get available topics
  • GET /api/v1/insights/sentiments - Get available sentiments
  • GET /api/v1/insights - Get analytics data

Monitoring

  • GET /api/v1/health - Check system health

📖 Code Examples Included

Languages Covered

  • Python: Complete examples for all major operations
  • JavaScript/Node.js: Promise-based and async/await examples
  • C#: .NET HttpClient examples
  • cURL: Command-line examples

Example Scenarios

  • File transcription workflow
  • Live transcription streaming
  • Sentiment analysis integration
  • Entity extraction
  • Topic analysis
  • Multi-language translation
  • Dashboard metrics
  • Health monitoring
  • CRM integration

✅ Complete API Coverage

All endpoints from the Swagger definition have been documented with:

  • ✓ Request/response examples
  • ✓ Parameter descriptions
  • ✓ Response schema details
  • ✓ Error handling
  • ✓ Code samples
  • ✓ Use cases
  • ✓ Integration guides
  • ✓ Best practices

🎓 Learning Path

Beginner

  1. Read Introduction
  2. Learn Authentication
  3. Try Speech-to-Text Getting Started

Intermediate

  1. Explore Speech-to-Text Features
  2. Add Sentiment Analysis
  3. Extract Topics

Advanced

  1. Implement Live Transcription
  2. Add Named Entity Recognition
  3. Use Translation
  4. Integrate Analytics

Operations

  1. Set up Health Monitoring
  2. Create alerts and dashboards
  3. Monitor performance metrics
  4. Plan capacity

🔍 How to Find Information

By Feature

  • Transcription: /docs/speech-to-text/
  • Real-time: /docs/live-speech/
  • Insights: /docs/audio-inteligence/
  • Analytics: /docs/insights
  • Operations: /docs/health-check

By Task

  • Set up API access: /docs/authentication
  • Send audio to transcribe: /docs/speech-to-text/getting-started
  • Stream live audio: /docs/live-speech/getting-started
  • Analyze conversations: /docs/audio-inteligence/sentiment
  • Monitor system: /docs/health-check
  • View metrics: /docs/insights

By Language

  • Python: See /docs/speech-to-text/getting-started or any feature page
  • JavaScript: See /docs/live-speech/getting-started for WebSocket examples
  • C#: See /docs/speech-to-text/getting-started for HttpClient examples
  • cURL: See any endpoint documentation

📝 Code Example Format

All code examples follow a consistent format:

[Language Name]
- Complete, runnable code
- Clear variable names
- Comments explaining key steps
- Error handling
- Best practices demonstrated

🚀 Next Steps

  1. Start with API Access: Follow Authentication
  2. Try the API: Use examples from getting started guides
  3. Build Integration: Follow use case specific guides
  4. Monitor Operations: Set up Health Check alerts
  5. Optimize: Use Insights to track performance

❓ Support

For additional help:

  • Review code examples in each section
  • Check troubleshooting sections
  • Consult best practices
  • Contact support team

📊 Documentation Statistics

  • Total Pages: 12+
  • Code Examples: 50+
  • Supported Languages: 4 (Python, JavaScript, C#, cURL)
  • API Endpoints: 9 major endpoints
  • Audio Intelligence Features: 5 (Sentiment, Topic, NER, Summarization, Translation)
  • Use Cases Documented: 20+