mirror of
https://github.com/HifiExperiments/overte.git
synced 2025-08-09 20:36:24 +02:00
include codec in audio stream packets so that each side can discard packets that don't match
This commit is contained in:
parent
dc6ab167e4
commit
20824f038c
12 changed files with 84 additions and 19 deletions
|
@ -768,28 +768,39 @@ void AudioMixer::broadcastMixes() {
|
||||||
|
|
||||||
std::unique_ptr<NLPacket> mixPacket;
|
std::unique_ptr<NLPacket> mixPacket;
|
||||||
|
|
||||||
|
const int MAX_CODEC_NAME = 30; // way over estimate
|
||||||
|
|
||||||
if (mixHasAudio) {
|
if (mixHasAudio) {
|
||||||
int mixPacketBytes = sizeof(quint16) + AudioConstants::NETWORK_FRAME_BYTES_STEREO;
|
int mixPacketBytes = sizeof(quint16) + MAX_CODEC_NAME+ AudioConstants::NETWORK_FRAME_BYTES_STEREO;
|
||||||
mixPacket = NLPacket::create(PacketType::MixedAudio, mixPacketBytes);
|
mixPacket = NLPacket::create(PacketType::MixedAudio, mixPacketBytes);
|
||||||
|
|
||||||
// pack sequence number
|
// pack sequence number
|
||||||
quint16 sequence = nodeData->getOutgoingSequenceNumber();
|
quint16 sequence = nodeData->getOutgoingSequenceNumber();
|
||||||
mixPacket->writePrimitive(sequence);
|
mixPacket->writePrimitive(sequence);
|
||||||
|
|
||||||
|
// write the codec
|
||||||
|
QString codecInPacket = nodeData->getCodecName();
|
||||||
|
mixPacket->writeString(codecInPacket);
|
||||||
|
|
||||||
QByteArray decodedBuffer(reinterpret_cast<char*>(_clampedSamples), AudioConstants::NETWORK_FRAME_BYTES_STEREO);
|
QByteArray decodedBuffer(reinterpret_cast<char*>(_clampedSamples), AudioConstants::NETWORK_FRAME_BYTES_STEREO);
|
||||||
QByteArray encodedBuffer;
|
QByteArray encodedBuffer;
|
||||||
nodeData->encode(decodedBuffer, encodedBuffer);
|
nodeData->encode(decodedBuffer, encodedBuffer);
|
||||||
|
|
||||||
// pack mixed audio samples
|
// pack mixed audio samples
|
||||||
mixPacket->write(encodedBuffer.constData(), encodedBuffer.size());
|
mixPacket->write(encodedBuffer.constData(), encodedBuffer.size());
|
||||||
} else {
|
}
|
||||||
int silentPacketBytes = sizeof(quint16) + sizeof(quint16);
|
else {
|
||||||
|
int silentPacketBytes = sizeof(quint16) + sizeof(quint16) + MAX_CODEC_NAME;
|
||||||
mixPacket = NLPacket::create(PacketType::SilentAudioFrame, silentPacketBytes);
|
mixPacket = NLPacket::create(PacketType::SilentAudioFrame, silentPacketBytes);
|
||||||
|
|
||||||
// pack sequence number
|
// pack sequence number
|
||||||
quint16 sequence = nodeData->getOutgoingSequenceNumber();
|
quint16 sequence = nodeData->getOutgoingSequenceNumber();
|
||||||
mixPacket->writePrimitive(sequence);
|
mixPacket->writePrimitive(sequence);
|
||||||
|
|
||||||
|
// write the codec
|
||||||
|
QString codecInPacket = nodeData->getCodecName();
|
||||||
|
mixPacket->writeString(codecInPacket);
|
||||||
|
|
||||||
// pack number of silent audio samples
|
// pack number of silent audio samples
|
||||||
quint16 numSilentSamples = AudioConstants::NETWORK_FRAME_SAMPLES_STEREO;
|
quint16 numSilentSamples = AudioConstants::NETWORK_FRAME_SAMPLES_STEREO;
|
||||||
mixPacket->writePrimitive(numSilentSamples);
|
mixPacket->writePrimitive(numSilentSamples);
|
||||||
|
|
|
@ -128,7 +128,6 @@ int AudioMixerClientData::parseData(ReceivedMessage& message) {
|
||||||
isMicStream = true;
|
isMicStream = true;
|
||||||
} else if (packetType == PacketType::InjectAudio) {
|
} else if (packetType == PacketType::InjectAudio) {
|
||||||
// this is injected audio
|
// this is injected audio
|
||||||
|
|
||||||
// grab the stream identifier for this injected audio
|
// grab the stream identifier for this injected audio
|
||||||
message.seek(sizeof(quint16));
|
message.seek(sizeof(quint16));
|
||||||
QUuid streamIdentifier = QUuid::fromRfc4122(message.readWithoutCopy(NUM_BYTES_RFC4122_UUID));
|
QUuid streamIdentifier = QUuid::fromRfc4122(message.readWithoutCopy(NUM_BYTES_RFC4122_UUID));
|
||||||
|
@ -167,6 +166,7 @@ int AudioMixerClientData::parseData(ReceivedMessage& message) {
|
||||||
|
|
||||||
// check the overflow count before we parse data
|
// check the overflow count before we parse data
|
||||||
auto overflowBefore = matchingStream->getOverflowCount();
|
auto overflowBefore = matchingStream->getOverflowCount();
|
||||||
|
|
||||||
auto parseResult = matchingStream->parseData(message);
|
auto parseResult = matchingStream->parseData(message);
|
||||||
|
|
||||||
if (matchingStream->getOverflowCount() > overflowBefore) {
|
if (matchingStream->getOverflowCount() > overflowBefore) {
|
||||||
|
|
|
@ -78,6 +78,9 @@ public:
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
QString getCodecName() { return _selectedCodecName; }
|
||||||
|
|
||||||
|
|
||||||
signals:
|
signals:
|
||||||
void injectorStreamFinished(const QUuid& streamIdentifier);
|
void injectorStreamFinished(const QUuid& streamIdentifier);
|
||||||
|
|
||||||
|
|
|
@ -834,7 +834,7 @@ void AudioClient::handleAudioInput() {
|
||||||
encodedBuffer = decocedBuffer;
|
encodedBuffer = decocedBuffer;
|
||||||
}
|
}
|
||||||
|
|
||||||
emitAudioPacket(encodedBuffer.constData(), encodedBuffer.size(), _outgoingAvatarAudioSequenceNumber, audioTransform, packetType);
|
emitAudioPacket(encodedBuffer.constData(), encodedBuffer.size(), _outgoingAvatarAudioSequenceNumber, audioTransform, packetType, _selectedCodecName);
|
||||||
_stats.sentPacket();
|
_stats.sentPacket();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -852,7 +852,7 @@ void AudioClient::handleRecordedAudioInput(const QByteArray& audio) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// FIXME check a flag to see if we should echo audio?
|
// FIXME check a flag to see if we should echo audio?
|
||||||
emitAudioPacket(encodedBuffer.data(), encodedBuffer.size(), _outgoingAvatarAudioSequenceNumber, audioTransform, PacketType::MicrophoneAudioWithEcho);
|
emitAudioPacket(encodedBuffer.data(), encodedBuffer.size(), _outgoingAvatarAudioSequenceNumber, audioTransform, PacketType::MicrophoneAudioWithEcho, _selectedCodecName);
|
||||||
}
|
}
|
||||||
|
|
||||||
void AudioClient::mixLocalAudioInjectors(int16_t* inputBuffer) {
|
void AudioClient::mixLocalAudioInjectors(int16_t* inputBuffer) {
|
||||||
|
@ -1015,7 +1015,6 @@ bool AudioClient::outputLocalInjector(bool isStereo, AudioInjector* injector) {
|
||||||
// no reason to lock access to the vector of injectors.
|
// no reason to lock access to the vector of injectors.
|
||||||
if (!_activeLocalAudioInjectors.contains(injector)) {
|
if (!_activeLocalAudioInjectors.contains(injector)) {
|
||||||
qDebug() << "adding new injector";
|
qDebug() << "adding new injector";
|
||||||
|
|
||||||
_activeLocalAudioInjectors.append(injector);
|
_activeLocalAudioInjectors.append(injector);
|
||||||
} else {
|
} else {
|
||||||
qDebug() << "injector exists in active list already";
|
qDebug() << "injector exists in active list already";
|
||||||
|
|
|
@ -19,7 +19,8 @@
|
||||||
|
|
||||||
#include "AudioConstants.h"
|
#include "AudioConstants.h"
|
||||||
|
|
||||||
void AbstractAudioInterface::emitAudioPacket(const void* audioData, size_t bytes, quint16& sequenceNumber, const Transform& transform, PacketType packetType) {
|
void AbstractAudioInterface::emitAudioPacket(const void* audioData, size_t bytes, quint16& sequenceNumber,
|
||||||
|
const Transform& transform, PacketType packetType, QString codecName) {
|
||||||
static std::mutex _mutex;
|
static std::mutex _mutex;
|
||||||
using Locker = std::unique_lock<std::mutex>;
|
using Locker = std::unique_lock<std::mutex>;
|
||||||
auto nodeList = DependencyManager::get<NodeList>();
|
auto nodeList = DependencyManager::get<NodeList>();
|
||||||
|
@ -27,10 +28,19 @@ void AbstractAudioInterface::emitAudioPacket(const void* audioData, size_t bytes
|
||||||
if (audioMixer && audioMixer->getActiveSocket()) {
|
if (audioMixer && audioMixer->getActiveSocket()) {
|
||||||
Locker lock(_mutex);
|
Locker lock(_mutex);
|
||||||
auto audioPacket = NLPacket::create(packetType);
|
auto audioPacket = NLPacket::create(packetType);
|
||||||
|
|
||||||
|
// FIXME - this is not a good way to determine stereoness with codecs....
|
||||||
quint8 isStereo = bytes == AudioConstants::NETWORK_FRAME_BYTES_STEREO ? 1 : 0;
|
quint8 isStereo = bytes == AudioConstants::NETWORK_FRAME_BYTES_STEREO ? 1 : 0;
|
||||||
|
|
||||||
// write sequence number
|
// write sequence number
|
||||||
audioPacket->writePrimitive(sequenceNumber++);
|
auto sequence = sequenceNumber++;
|
||||||
|
audioPacket->writePrimitive(sequence);
|
||||||
|
|
||||||
|
// write the codec - don't include this for injected audio
|
||||||
|
if (packetType != PacketType::InjectAudio) {
|
||||||
|
auto stringSize = audioPacket->writeString(codecName);
|
||||||
|
}
|
||||||
|
|
||||||
if (packetType == PacketType::SilentAudioFrame) {
|
if (packetType == PacketType::SilentAudioFrame) {
|
||||||
// pack num silent samples
|
// pack num silent samples
|
||||||
quint16 numSilentSamples = isStereo ?
|
quint16 numSilentSamples = isStereo ?
|
||||||
|
@ -49,8 +59,8 @@ void AbstractAudioInterface::emitAudioPacket(const void* audioData, size_t bytes
|
||||||
|
|
||||||
if (audioPacket->getType() != PacketType::SilentAudioFrame) {
|
if (audioPacket->getType() != PacketType::SilentAudioFrame) {
|
||||||
// audio samples have already been packed (written to networkAudioSamples)
|
// audio samples have already been packed (written to networkAudioSamples)
|
||||||
audioPacket->setPayloadSize(audioPacket->getPayloadSize() + bytes);
|
int leadingBytes = audioPacket->getPayloadSize();
|
||||||
static const int leadingBytes = sizeof(quint16) + sizeof(glm::vec3) + sizeof(glm::quat) + sizeof(quint8);
|
audioPacket->setPayloadSize(leadingBytes + bytes);
|
||||||
memcpy(audioPacket->getPayload() + leadingBytes, audioData, bytes);
|
memcpy(audioPacket->getPayload() + leadingBytes, audioData, bytes);
|
||||||
}
|
}
|
||||||
nodeList->flagTimeForConnectionStep(LimitedNodeList::ConnectionStep::SendAudioPacket);
|
nodeList->flagTimeForConnectionStep(LimitedNodeList::ConnectionStep::SendAudioPacket);
|
||||||
|
|
|
@ -28,7 +28,8 @@ class AbstractAudioInterface : public QObject {
|
||||||
public:
|
public:
|
||||||
AbstractAudioInterface(QObject* parent = 0) : QObject(parent) {};
|
AbstractAudioInterface(QObject* parent = 0) : QObject(parent) {};
|
||||||
|
|
||||||
static void emitAudioPacket(const void* audioData, size_t bytes, quint16& sequenceNumber, const Transform& transform, PacketType packetType);
|
static void emitAudioPacket(const void* audioData, size_t bytes, quint16& sequenceNumber, const Transform& transform,
|
||||||
|
PacketType packetType, QString codecName = QString(""));
|
||||||
|
|
||||||
public slots:
|
public slots:
|
||||||
virtual bool outputLocalInjector(bool isStereo, AudioInjector* injector) = 0;
|
virtual bool outputLocalInjector(bool isStereo, AudioInjector* injector) = 0;
|
||||||
|
|
|
@ -218,6 +218,14 @@ const uchar MAX_INJECTOR_VOLUME = 0xFF;
|
||||||
static const int64_t NEXT_FRAME_DELTA_ERROR_OR_FINISHED = -1;
|
static const int64_t NEXT_FRAME_DELTA_ERROR_OR_FINISHED = -1;
|
||||||
static const int64_t NEXT_FRAME_DELTA_IMMEDIATELY = 0;
|
static const int64_t NEXT_FRAME_DELTA_IMMEDIATELY = 0;
|
||||||
|
|
||||||
|
qint64 writeStringToStream(const QString& string, QDataStream& stream) {
|
||||||
|
QByteArray data = string.toUtf8();
|
||||||
|
uint32_t length = data.length();
|
||||||
|
stream << static_cast<quint32>(length);
|
||||||
|
stream << data;
|
||||||
|
return length + sizeof(uint32_t);
|
||||||
|
}
|
||||||
|
|
||||||
int64_t AudioInjector::injectNextFrame() {
|
int64_t AudioInjector::injectNextFrame() {
|
||||||
if (stateHas(AudioInjectorState::NetworkInjectionFinished)) {
|
if (stateHas(AudioInjectorState::NetworkInjectionFinished)) {
|
||||||
qDebug() << "AudioInjector::injectNextFrame called but AudioInjector has finished and was not restarted. Returning.";
|
qDebug() << "AudioInjector::injectNextFrame called but AudioInjector has finished and was not restarted. Returning.";
|
||||||
|
@ -264,6 +272,10 @@ int64_t AudioInjector::injectNextFrame() {
|
||||||
// pack some placeholder sequence number for now
|
// pack some placeholder sequence number for now
|
||||||
audioPacketStream << (quint16) 0;
|
audioPacketStream << (quint16) 0;
|
||||||
|
|
||||||
|
// pack some placeholder sequence number for now
|
||||||
|
//QString noCodecForInjectors("");
|
||||||
|
//writeStringToStream(noCodecForInjectors, audioPacketStream);
|
||||||
|
|
||||||
// pack stream identifier (a generated UUID)
|
// pack stream identifier (a generated UUID)
|
||||||
audioPacketStream << QUuid::createUuid();
|
audioPacketStream << QUuid::createUuid();
|
||||||
|
|
||||||
|
|
|
@ -58,6 +58,7 @@ void InboundAudioStream::reset() {
|
||||||
_isStarved = true;
|
_isStarved = true;
|
||||||
_hasStarted = false;
|
_hasStarted = false;
|
||||||
resetStats();
|
resetStats();
|
||||||
|
//cleanupCodec(); // FIXME???
|
||||||
}
|
}
|
||||||
|
|
||||||
void InboundAudioStream::resetStats() {
|
void InboundAudioStream::resetStats() {
|
||||||
|
@ -99,12 +100,17 @@ void InboundAudioStream::perSecondCallbackForUpdatingStats() {
|
||||||
}
|
}
|
||||||
|
|
||||||
int InboundAudioStream::parseData(ReceivedMessage& message) {
|
int InboundAudioStream::parseData(ReceivedMessage& message) {
|
||||||
|
PacketType packetType = message.getType();
|
||||||
|
|
||||||
// parse sequence number and track it
|
// parse sequence number and track it
|
||||||
quint16 sequence;
|
quint16 sequence;
|
||||||
message.readPrimitive(&sequence);
|
message.readPrimitive(&sequence);
|
||||||
SequenceNumberStats::ArrivalInfo arrivalInfo = _incomingSequenceNumberStats.sequenceNumberReceived(sequence,
|
SequenceNumberStats::ArrivalInfo arrivalInfo = _incomingSequenceNumberStats.sequenceNumberReceived(sequence,
|
||||||
message.getSourceID());
|
message.getSourceID());
|
||||||
|
QString codecInPacket("");
|
||||||
|
if (packetType != PacketType::InjectAudio) {
|
||||||
|
codecInPacket = message.readString();
|
||||||
|
}
|
||||||
|
|
||||||
packetReceivedUpdateTimingStats();
|
packetReceivedUpdateTimingStats();
|
||||||
|
|
||||||
|
@ -112,9 +118,10 @@ int InboundAudioStream::parseData(ReceivedMessage& message) {
|
||||||
|
|
||||||
// parse the info after the seq number and before the audio data (the stream properties)
|
// parse the info after the seq number and before the audio data (the stream properties)
|
||||||
int prePropertyPosition = message.getPosition();
|
int prePropertyPosition = message.getPosition();
|
||||||
int propertyBytes = parseStreamProperties(message.getType(), message.readWithoutCopy(message.getBytesLeftToRead()), networkSamples);
|
auto afterHeader = message.readWithoutCopy(message.getBytesLeftToRead());
|
||||||
|
int propertyBytes = parseStreamProperties(message.getType(), afterHeader, networkSamples);
|
||||||
message.seek(prePropertyPosition + propertyBytes);
|
message.seek(prePropertyPosition + propertyBytes);
|
||||||
|
|
||||||
// handle this packet based on its arrival status.
|
// handle this packet based on its arrival status.
|
||||||
switch (arrivalInfo._status) {
|
switch (arrivalInfo._status) {
|
||||||
case SequenceNumberStats::Early: {
|
case SequenceNumberStats::Early: {
|
||||||
|
@ -129,9 +136,19 @@ int InboundAudioStream::parseData(ReceivedMessage& message) {
|
||||||
case SequenceNumberStats::OnTime: {
|
case SequenceNumberStats::OnTime: {
|
||||||
// Packet is on time; parse its data to the ringbuffer
|
// Packet is on time; parse its data to the ringbuffer
|
||||||
if (message.getType() == PacketType::SilentAudioFrame) {
|
if (message.getType() == PacketType::SilentAudioFrame) {
|
||||||
|
// FIXME - do some codecs need to know about these silen frames?
|
||||||
writeDroppableSilentSamples(networkSamples);
|
writeDroppableSilentSamples(networkSamples);
|
||||||
} else {
|
} else {
|
||||||
parseAudioData(message.getType(), message.readWithoutCopy(message.getBytesLeftToRead()));
|
// note: PCM and no codec are identical
|
||||||
|
bool selectedPCM = _selectedCodecName == "pcm" || _selectedCodecName == "";
|
||||||
|
bool packetPCM = codecInPacket == "pcm" || codecInPacket == "";
|
||||||
|
if (codecInPacket == _selectedCodecName || (packetPCM && selectedPCM)) {
|
||||||
|
auto afterProperties = message.readWithoutCopy(message.getBytesLeftToRead());
|
||||||
|
parseAudioData(message.getType(), afterProperties);
|
||||||
|
} else {
|
||||||
|
qDebug() << __FUNCTION__ << "codec mismatch: expected" << _selectedCodecName << "got" << codecInPacket << "writing silence";
|
||||||
|
writeDroppableSilentSamples(networkSamples);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
|
@ -33,6 +33,7 @@ const uchar MAX_INJECTOR_VOLUME = 255;
|
||||||
int InjectedAudioStream::parseStreamProperties(PacketType type,
|
int InjectedAudioStream::parseStreamProperties(PacketType type,
|
||||||
const QByteArray& packetAfterSeqNum,
|
const QByteArray& packetAfterSeqNum,
|
||||||
int& numAudioSamples) {
|
int& numAudioSamples) {
|
||||||
|
|
||||||
// setup a data stream to read from this packet
|
// setup a data stream to read from this packet
|
||||||
QDataStream packetStream(packetAfterSeqNum);
|
QDataStream packetStream(packetAfterSeqNum);
|
||||||
|
|
||||||
|
|
|
@ -154,8 +154,8 @@ qint64 BasePacket::writeString(const QString& string) {
|
||||||
QByteArray data = string.toUtf8();
|
QByteArray data = string.toUtf8();
|
||||||
uint32_t length = data.length();
|
uint32_t length = data.length();
|
||||||
writePrimitive(length);
|
writePrimitive(length);
|
||||||
writeData(data.constData(), data.length());
|
write(data.constData(), data.length());
|
||||||
seek(pos() + length);
|
//seek(pos() + length);
|
||||||
return length + sizeof(uint32_t);
|
return length + sizeof(uint32_t);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -176,7 +176,6 @@ bool BasePacket::reset() {
|
||||||
}
|
}
|
||||||
|
|
||||||
qint64 BasePacket::writeData(const char* data, qint64 maxSize) {
|
qint64 BasePacket::writeData(const char* data, qint64 maxSize) {
|
||||||
|
|
||||||
Q_ASSERT_X(maxSize <= bytesAvailableForWrite(), "BasePacket::writeData", "not enough space for write");
|
Q_ASSERT_X(maxSize <= bytesAvailableForWrite(), "BasePacket::writeData", "not enough space for write");
|
||||||
|
|
||||||
// make sure we have the space required to write this block
|
// make sure we have the space required to write this block
|
||||||
|
|
|
@ -72,6 +72,13 @@ PacketVersion versionForPacketType(PacketType packetType) {
|
||||||
case PacketType::DomainServerAddedNode:
|
case PacketType::DomainServerAddedNode:
|
||||||
return static_cast<PacketVersion>(DomainServerAddedNodeVersion::PermissionsGrid);
|
return static_cast<PacketVersion>(DomainServerAddedNodeVersion::PermissionsGrid);
|
||||||
|
|
||||||
|
case PacketType::MixedAudio:
|
||||||
|
case PacketType::SilentAudioFrame:
|
||||||
|
case PacketType::InjectAudio:
|
||||||
|
case PacketType::MicrophoneAudioNoEcho:
|
||||||
|
case PacketType::MicrophoneAudioWithEcho:
|
||||||
|
return static_cast<PacketVersion>(AudioVersion::CodecNameInAudioPackets);
|
||||||
|
|
||||||
default:
|
default:
|
||||||
return 17;
|
return 17;
|
||||||
}
|
}
|
||||||
|
|
|
@ -213,4 +213,9 @@ enum class DomainListVersion : PacketVersion {
|
||||||
PermissionsGrid
|
PermissionsGrid
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum class AudioVersion : PacketVersion {
|
||||||
|
HasCompressedAudio = 17,
|
||||||
|
CodecNameInAudioPackets
|
||||||
|
};
|
||||||
|
|
||||||
#endif // hifi_PacketHeaders_h
|
#endif // hifi_PacketHeaders_h
|
||||||
|
|
Loading…
Reference in a new issue