haikuwebkit/Source/WebCore/Modules/speech/SpeechRecognition.cpp

209 lines
7.3 KiB
C++

/*
* Copyright (C) 2020 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS''
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
* THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "config.h"
#include "SpeechRecognition.h"
#include "ClientOrigin.h"
#include "Document.h"
#include "EventNames.h"
#include "Page.h"
#include "SpeechRecognitionError.h"
#include "SpeechRecognitionErrorEvent.h"
#include "SpeechRecognitionEvent.h"
#include "SpeechRecognitionResultData.h"
#include "SpeechRecognitionResultList.h"
#include <wtf/IsoMallocInlines.h>
namespace WebCore {
WTF_MAKE_ISO_ALLOCATED_IMPL(SpeechRecognition);
Ref<SpeechRecognition> SpeechRecognition::create(Document& document)
{
auto recognition = adoptRef(*new SpeechRecognition(document));
recognition->suspendIfNeeded();
return recognition;
}
SpeechRecognition::SpeechRecognition(Document& document)
: ActiveDOMObject(document)
{
if (auto* page = document.page()) {
m_connection = makeRefPtr(page->speechRecognitionConnection());
m_connection->registerClient(*this);
}
}
void SpeechRecognition::suspend(ReasonForSuspension)
{
abortRecognition();
}
ExceptionOr<void> SpeechRecognition::startRecognition()
{
if (m_state != State::Inactive)
return Exception { InvalidStateError, "Recognition is being started or already started"_s };
if (!m_connection)
return Exception { UnknownError, "Recognition does not have a valid connection"_s };
auto& document = downcast<Document>(*scriptExecutionContext());
auto* frame = document.frame();
if (!frame)
return Exception { UnknownError, "Recognition is not in a valid frame"_s };
auto optionalFrameIdentifier = document.frameID();
auto frameIdentifier = optionalFrameIdentifier ? *optionalFrameIdentifier : FrameIdentifier { };
m_connection->start(identifier(), m_lang, m_continuous, m_interimResults, m_maxAlternatives, ClientOrigin { document.topOrigin().data(), document.securityOrigin().data() }, frameIdentifier);
m_state = State::Starting;
return { };
}
void SpeechRecognition::stopRecognition()
{
if (m_state == State::Inactive || m_state == State::Stopping || m_state == State::Aborting)
return;
m_connection->stop(identifier());
m_state = State::Stopping;
}
void SpeechRecognition::abortRecognition()
{
if (m_state == State::Inactive || m_state == State::Aborting)
return;
m_connection->abort(identifier());
m_state = State::Aborting;
}
const char* SpeechRecognition::activeDOMObjectName() const
{
return "SpeechRecognition";
}
void SpeechRecognition::stop()
{
abortRecognition();
if (!m_connection)
return;
m_connection->unregisterClient(*this);
auto& document = downcast<Document>(*scriptExecutionContext());
document.setActiveSpeechRecognition(nullptr);
}
void SpeechRecognition::didStart()
{
if (m_state == State::Starting)
m_state = State::Running;
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().startEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStartCapturingAudio()
{
auto& document = downcast<Document>(*scriptExecutionContext());
document.setActiveSpeechRecognition(this);
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().audiostartEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStartCapturingSound()
{
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().soundstartEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStartCapturingSpeech()
{
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().speechstartEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStopCapturingSpeech()
{
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().speechendEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStopCapturingSound()
{
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().soundendEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didStopCapturingAudio()
{
auto& document = downcast<Document>(*scriptExecutionContext());
document.setActiveSpeechRecognition(nullptr);
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().audioendEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
void SpeechRecognition::didFindNoMatch()
{
queueTaskToDispatchEvent(*this, TaskSource::Speech, SpeechRecognitionEvent::create(eventNames().nomatchEvent, 0, nullptr));
}
void SpeechRecognition::didReceiveResult(Vector<SpeechRecognitionResultData>&& resultDatas)
{
Vector<Ref<SpeechRecognitionResult>> allResults;
allResults.reserveCapacity(m_finalResults.size() + resultDatas.size());
allResults.appendVector(m_finalResults);
auto firstChangedIndex = allResults.size();
for (auto resultData : resultDatas) {
auto alternatives = WTF::map(resultData.alternatives, [](auto& alternativeData) {
return SpeechRecognitionAlternative::create(WTFMove(alternativeData.transcript), alternativeData.confidence);
});
auto newResult = SpeechRecognitionResult::create(WTFMove(alternatives), resultData.isFinal);
if (resultData.isFinal)
m_finalResults.append(newResult);
allResults.append(WTFMove(newResult));
}
auto resultList = SpeechRecognitionResultList::create(WTFMove(allResults));
queueTaskToDispatchEvent(*this, TaskSource::Speech, SpeechRecognitionEvent::create(eventNames().resultEvent, firstChangedIndex, WTFMove(resultList)));
}
void SpeechRecognition::didError(const SpeechRecognitionError& error)
{
m_finalResults.clear();
m_state = State::Inactive;
queueTaskToDispatchEvent(*this, TaskSource::Speech, SpeechRecognitionErrorEvent::create(eventNames().errorEvent, error.type, error.message));
}
void SpeechRecognition::didEnd()
{
m_finalResults.clear();
m_state = State::Inactive;
queueTaskToDispatchEvent(*this, TaskSource::Speech, Event::create(eventNames().endEvent, Event::CanBubble::No, Event::IsCancelable::No));
}
} // namespace WebCore