Talk Analytics API, gRPC: TalkService.Get
- gRPC request
- GetTalkRequest
- GetTalkResponse
- Talk
- Field
- Transcription
- Phrase
- PhraseText
- Word
- PhraseStatistics
- UtteranceStatistics
- AudioSegmentBoundaries
- DescriptiveStatistics
- Quantile
- RecognitionClassifierResult
- PhraseHighlight
- RecognitionClassifierLabel
- AlgorithmMetadata
- Error
- SpeechStatistics
- SilenceStatistics
- InterruptsStatistics
- InterruptsEvaluation
- ConversationStatistics
- SpeakerStatistics
- Points
- Quiz
- TextClassifiers
- ClassificationResult
- ClassifierStatistics
- Histogram
- Summarization
- SummarizationStatement
- SummarizationField
- TalkState
- AlgorithmProcessingInfo
rpc for bulk get
gRPC request
rpc Get (GetTalkRequest) returns (GetTalkResponse)
GetTalkRequest
{
"organization_id": "string",
"space_id": "string",
"connection_id": "string",
"project_id": "string",
"talk_ids": [
"string"
],
"results_mask": "google.protobuf.FieldMask"
}
Field |
Description |
organization_id |
string id of organization |
space_id |
string id of space |
connection_id |
string id of connection to search data |
project_id |
string id of project to search data |
talk_ids[] |
string ids of talks to return. Requesting too many talks may result in "message exceeds maximum size" error. |
results_mask |
All types of analysis will be returned if not set. |
GetTalkResponse
{
"talk": [
{
"id": "string",
"organization_id": "string",
"space_id": "string",
"connection_id": "string",
"project_ids": [
"string"
],
"created_by": "string",
"created_at": "google.protobuf.Timestamp",
"modified_by": "string",
"modified_at": "google.protobuf.Timestamp",
"talk_fields": [
{
"name": "string",
"value": "string",
"type": "FieldType"
}
],
"transcription": {
"phrases": [
{
"channel_number": "int64",
"start_time_ms": "int64",
"end_time_ms": "int64",
"phrase": {
"text": "string",
"language": "string",
"normalized_text": "string",
"words": [
{
"word": "string",
"start_time_ms": "int64",
"end_time_ms": "int64"
}
]
},
"statistics": {
"statistics": {
"speaker_tag": "string",
"speech_boundaries": {
"start_time_ms": "int64",
"end_time_ms": "int64",
"duration_seconds": "int64"
},
"total_speech_ms": "int64",
"speech_ratio": "double",
"total_silence_ms": "int64",
"silence_ratio": "double",
"words_count": "int64",
"letters_count": "int64",
"words_per_second": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
},
"letters_per_second": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
}
}
},
"classifiers": [
{
"start_time_ms": "int64",
"end_time_ms": "int64",
"classifier": "string",
"highlights": [
{
"text": "string",
"offset": "int64",
"count": "int64"
}
],
"labels": [
{
"label": "string",
"confidence": "double"
}
]
}
]
}
],
"algorithms_metadata": [
{
"created_task_date": "google.protobuf.Timestamp",
"completed_task_date": "google.protobuf.Timestamp",
"error": {
"code": "string",
"message": "string"
},
"trace_id": "string",
"name": "string"
}
]
},
"speech_statistics": {
"total_simultaneous_speech_duration_seconds": "int64",
"total_simultaneous_speech_duration_ms": "int64",
"total_simultaneous_speech_ratio": "double",
"simultaneous_speech_duration_estimation": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
}
},
"silence_statistics": {
"total_simultaneous_silence_duration_ms": "int64",
"total_simultaneous_silence_ratio": "double",
"simultaneous_silence_duration_estimation": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
},
"total_simultaneous_silence_duration_seconds": "int64"
},
"interrupts_statistics": {
"speaker_interrupts": [
{
"speaker_tag": "string",
"interrupts_count": "int64",
"interrupts_duration_ms": "int64",
"interrupts": [
{
"start_time_ms": "int64",
"end_time_ms": "int64",
"duration_seconds": "int64"
}
],
"interrupts_duration_seconds": "int64"
}
]
},
"conversation_statistics": {
"conversation_boundaries": {
"start_time_ms": "int64",
"end_time_ms": "int64",
"duration_seconds": "int64"
},
"speaker_statistics": [
{
"speaker_tag": "string",
"complete_statistics": {
"speaker_tag": "string",
"speech_boundaries": {
"start_time_ms": "int64",
"end_time_ms": "int64",
"duration_seconds": "int64"
},
"total_speech_ms": "int64",
"speech_ratio": "double",
"total_silence_ms": "int64",
"silence_ratio": "double",
"words_count": "int64",
"letters_count": "int64",
"words_per_second": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
},
"letters_per_second": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
}
},
"words_per_utterance": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
},
"letters_per_utterance": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
},
"utterance_count": "int64",
"utterance_duration_estimation": {
"min": "double",
"max": "double",
"mean": "double",
"std": "double",
"quantiles": [
{
"level": "double",
"value": "double"
}
]
}
}
]
},
"points": {
"quiz": [
{
"request": "string",
"response": "google.protobuf.StringValue",
"id": "string"
}
]
},
"text_classifiers": {
"classification_result": [
{
"classifier": "string",
"classifier_statistics": [
{
"channel_number": "google.protobuf.Int64Value",
"total_count": "int64",
"histograms": [
{
"count_values": [
"int64"
]
}
]
}
]
}
]
},
"summarization": {
"statements": [
{
"field": {
"id": "string",
"name": "string",
"type": "SummarizationFieldType"
},
"response": [
"string"
]
}
]
},
"talk_state": {
"processing_state": "ProcessingState",
"algorithm_processing_infos": [
{
"algorithm": "Algorithm",
"processing_state": "ProcessingState"
}
]
}
}
]
}
Field |
Description |
talk[] |
Talk
Field |
Description |
id |
string talk id |
organization_id |
string |
space_id |
string |
connection_id |
string |
project_ids[] |
string |
created_by |
string audition info |
created_at |
|
modified_by |
string |
modified_at |
|
talk_fields[] |
key-value representation of talk fields with values |
transcription |
various ml analysis results |
speech_statistics |
|
silence_statistics |
|
interrupts_statistics |
|
conversation_statistics |
|
points |
|
text_classifiers |
|
summarization |
|
talk_state |
Field
connection field value
Field |
Description |
name |
string name of the field |
value |
string field value |
type |
enum FieldType field type
|
Transcription
Field |
Description |
phrases[] |
|
algorithms_metadata[] |
Their might be several algorithms that work on talk transcription. For example: speechkit and translator |
Phrase
Field |
Description |
channel_number |
int64 |
start_time_ms |
int64 |
end_time_ms |
int64 |
phrase |
|
statistics |
|
classifiers[] |
PhraseText
Field |
Description |
text |
string |
language |
string |
normalized_text |
string |
words[] |
Word
Field |
Description |
word |
string |
start_time_ms |
int64 |
end_time_ms |
int64 |
PhraseStatistics
Field |
Description |
statistics |
UtteranceStatistics
Field |
Description |
speaker_tag |
string |
speech_boundaries |
Audio segment boundaries |
total_speech_ms |
int64 Total speech duration |
speech_ratio |
double Speech ratio within audio segment |
total_silence_ms |
int64 Total silence duration |
silence_ratio |
double Silence ratio within audio segment |
words_count |
int64 Number of words in recognized speech |
letters_count |
int64 Number of letters in recognized speech |
words_per_second |
Descriptive statistics for words per second distribution |
letters_per_second |
Descriptive statistics for letters per second distribution |
AudioSegmentBoundaries
Field |
Description |
start_time_ms |
int64 Audio segment start time |
end_time_ms |
int64 Audio segment end time |
duration_seconds |
int64 Duration in seconds |
DescriptiveStatistics
Field |
Description |
min |
double Minimum observed value |
max |
double Maximum observed value |
mean |
double Estimated mean of distribution |
std |
double Estimated standard deviation of distribution |
quantiles[] |
List of evaluated quantiles |
Quantile
Field |
Description |
level |
double Quantile level in range (0, 1) |
value |
double Quantile value |
RecognitionClassifierResult
Field |
Description |
start_time_ms |
int64 Start time of the audio segment used for classification |
end_time_ms |
int64 End time of the audio segment used for classification |
classifier |
string Name of the triggered classifier |
highlights[] |
List of highlights, i.e. parts of phrase that determine the result of the classification |
labels[] |
Classifier predictions |
PhraseHighlight
Field |
Description |
text |
string Text transcription of the highlighted audio segment |
offset |
int64 offset in symbols from the beginning of whole phrase where highlight begins |
count |
int64 count of symbols in highlighted text |
RecognitionClassifierLabel
Field |
Description |
label |
string The label of the class predicted by the classifier |
confidence |
double The prediction confidence |
AlgorithmMetadata
Field |
Description |
created_task_date |
|
completed_task_date |
|
error |
|
trace_id |
string |
name |
string |
Error
Field |
Description |
code |
string |
message |
string |
SpeechStatistics
Field |
Description |
total_simultaneous_speech_duration_seconds |
int64 Total simultaneous speech duration in seconds |
total_simultaneous_speech_duration_ms |
int64 Total simultaneous speech duration in ms |
total_simultaneous_speech_ratio |
double Simultaneous speech ratio within audio segment |
simultaneous_speech_duration_estimation |
Descriptive statistics for simultaneous speech duration distribution |
SilenceStatistics
Field |
Description |
total_simultaneous_silence_duration_ms |
int64 |
total_simultaneous_silence_ratio |
double Simultaneous silence ratio within audio segment |
simultaneous_silence_duration_estimation |
Descriptive statistics for simultaneous silence duration distribution |
total_simultaneous_silence_duration_seconds |
int64 |
InterruptsStatistics
Field |
Description |
speaker_interrupts[] |
Interrupts description for every speaker |
InterruptsEvaluation
Field |
Description |
speaker_tag |
string Speaker tag |
interrupts_count |
int64 Number of interrupts made by the speaker |
interrupts_duration_ms |
int64 Total duration of all interrupts |
interrupts[] |
Boundaries for every interrupt |
interrupts_duration_seconds |
int64 Total duration of all interrupts in seconds |
ConversationStatistics
Field |
Description |
conversation_boundaries |
Audio segment boundaries |
speaker_statistics[] |
Average statistics for each speaker |
SpeakerStatistics
Field |
Description |
speaker_tag |
string Speaker tag |
complete_statistics |
analysis of all phrases in format of single utterance |
words_per_utterance |
Descriptive statistics for words per utterance distribution |
letters_per_utterance |
Descriptive statistics for letters per utterance distribution |
utterance_count |
int64 Number of utterances |
utterance_duration_estimation |
Descriptive statistics for utterance duration distribution |
Points
Field |
Description |
quiz[] |
Quiz
Field |
Description |
request |
string |
response |
|
id |
string |
TextClassifiers
Field |
Description |
classification_result[] |
ClassificationResult
Field |
Description |
classifier |
string Classifier name |
classifier_statistics[] |
Classifier statistics |
ClassifierStatistics
Field |
Description |
channel_number |
Channel number, null for whole talk |
total_count |
int64 classifier total count |
histograms[] |
Represents various histograms build on top of classifiers |
Histogram
Field |
Description |
count_values[] |
int64 histogram count values. For example: |
Summarization
Field |
Description |
statements[] |
SummarizationStatement
Field |
Description |
field |
|
response[] |
string |
SummarizationField
Field |
Description |
id |
string |
name |
string |
type |
enum SummarizationFieldType
|
TalkState
Field |
Description |
processing_state |
enum ProcessingState
|
algorithm_processing_infos[] |
AlgorithmProcessingInfo
Field |
Description |
algorithm |
enum Algorithm
|
processing_state |
enum ProcessingState
|