mistserver/src/output/output_progressive_mp4.cpp
Thulinma a084005644 Merge branch 'development' into LTS_development
# Conflicts:
#	src/output/output_progressive_mp4.cpp
2018-05-16 11:10:55 +02:00

1016 lines
37 KiB
C++

#include <mist/defines.h>
#include <mist/mp4.h>
#include <mist/mp4_generic.h>
#include <mist/checksum.h>
#include <mist/bitfields.h>
#include "output_progressive_mp4.h"
#include <inttypes.h>
namespace Mist{
OutProgressiveMP4::OutProgressiveMP4(Socket::Connection & conn) : HTTPOutput(conn){}
OutProgressiveMP4::~OutProgressiveMP4(){}
void OutProgressiveMP4::init(Util::Config * cfg){
HTTPOutput::init(cfg);
capa["name"] = "MP4";
capa["desc"] = "Enables HTTP protocol progressive streaming.";
capa["url_rel"] = "/$.mp4";
capa["url_match"][0u] = "/$.mp4";
capa["url_match"][1u] = "/$.3gp";
capa["codecs"][0u][0u].append("H264");
capa["codecs"][0u][0u].append("HEVC");
capa["codecs"][0u][1u].append("AAC");
capa["codecs"][0u][1u].append("MP3");
capa["codecs"][0u][1u].append("AC3");
capa["methods"][0u]["handler"] = "http";
capa["methods"][0u]["type"] = "html5/video/mp4";
capa["methods"][0u]["priority"] = 10ll;
//MP4 live is broken on Apple
capa["exceptions"]["live"] = JSON::fromString("[[\"blacklist\",[\"iPad\",\"iPhone\",\"iPod\",\"Safari\"]], [\"whitelist\",[\"Chrome\",\"Chromium\"]]]");
}
uint64_t OutProgressiveMP4::estimateFileSize(){
uint64_t retVal = 0;
for (std::set<unsigned long>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
for (std::deque<unsigned long>::iterator keyIt = myMeta.tracks[*it].keySizes.begin(); keyIt != myMeta.tracks[*it].keySizes.end(); keyIt++){
retVal += *keyIt;
}
}
return retVal * 1.1;
}
uint64_t OutProgressiveMP4::mp4HeaderSize(uint64_t & fileSize, int fragmented){
bool useLargeBoxes = !fragmented && (estimateFileSize() > 0xFFFFFFFFull);
uint64_t res = 36 // FTYP Box
+ 8 //MOOV box
+ 108; //MVHD Box
uint64_t firstms = 0xFFFFFFFFFFFFFFFFull;
for (std::set<unsigned long>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
if (myMeta.tracks[*it].firstms < firstms){
firstms = myMeta.tracks[*it].firstms;
}
}
for (std::set<unsigned long>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
DTSC::Track & thisTrack = myMeta.tracks[*it];
uint64_t tmpRes = 0;
uint64_t partCount = thisTrack.parts.size();
tmpRes += 8 //TRAK Box
+ 92 //TKHD Box
+ 36 //EDTS Box
+ 8 //MDIA Box
+ 32 //MDHD Box
+ 33 + thisTrack.getIdentifier().size() // HDLR Box
+ 8 //MINF Box
+ 36 //DINF Box
+ 8; // STBL Box
if (thisTrack.firstms != firstms){
tmpRes += 12;// EDTS entry extra
}
//These boxes are empty when generating fragmented output
tmpRes += 20 + (fragmented ? 0 : (partCount * 4));//STSZ
tmpRes += 16 + (fragmented ? 0 : (partCount * (useLargeBoxes ? 8 : 4)));//STCO
tmpRes += 16 + (fragmented ? 0 : (1 * 12));//STSC <-- Currently 1 entry, but might become more complex in near future
//Type-specific boxes
if (thisTrack.type == "video"){
tmpRes += 20//VMHD Box
+ 16 //STSD
+ 86 //AVC1
+ 8 + thisTrack.init.size();//avcC
if (!fragmented){
tmpRes += 16 + (thisTrack.keys.size() * 4);//STSS
}
}
if (thisTrack.type == "audio"){
tmpRes += 16//SMHD Box
+ 16//STSD
+ 36//MP4A
+ 35;
if (thisTrack.init.size()){
tmpRes += 2 + thisTrack.init.size();//ESDS
}
}
if (thisTrack.type == "meta"){
tmpRes += 12 //NMHD Box
+ 16//STSD
+ 64; //tx3g Box
}
if (!fragmented){
//Unfortunately, for our STTS and CTTS boxes, we need to loop through all parts of the track
uint64_t sttsCount = 1;
uint64_t prevDur = thisTrack.parts[0].getDuration();
uint64_t prevOffset = thisTrack.parts[0].getOffset();
uint64_t cttsCount = 1;
fileSize += thisTrack.parts[0].getSize();
for (unsigned int part = 1; part < partCount; ++part){
uint64_t partDur = thisTrack.parts[part].getDuration();
uint64_t partOffset = thisTrack.parts[part].getOffset();
uint64_t partSize = thisTrack.parts[part].getSize();
if (prevDur != partDur){
prevDur = partDur;
++sttsCount;
}
if (partOffset != prevOffset){
prevOffset = partOffset;
++cttsCount;
}
fileSize += partSize;
}
if (cttsCount == 1 && ! prevOffset){
cttsCount = 0;
}
tmpRes += 16 + (sttsCount * 8);//STTS
if (cttsCount){
tmpRes += 16 + (cttsCount * 8);//CTTS
}
}else{
tmpRes += 16;//empty STTS, no CTTS
}
res += tmpRes;
}
if (fragmented){
res += 8 + (selectedTracks.size() * 32);//Mvex + trex boxes
res += 1; //Horrible horrible length fix;
}else{
res += 8; //mdat beginning
}
fileSize += res;
MEDIUM_MSG("H size %llu, file: %llu", res, fileSize);
return res;
}
///\todo This function does not indicate errors anywhere... maybe fix this...
std::string OutProgressiveMP4::DTSCMeta2MP4Header(uint64_t & size, int fragmented){
if (myMeta.live){
needsLookAhead = 420;
}
//Make sure we have a proper being value for the size...
size = 0;
//Stores the result of the function
std::stringstream header;
//Determines whether the outputfile is larger than 4GB, in which case we need to use 64-bit boxes for offsets
bool useLargeBoxes = !fragmented && (estimateFileSize() > 0xFFFFFFFFull);
//Keeps track of the total size of the mdat box
uint64_t mdatSize = 0;
//Start actually creating the header
//MP4 Files always start with an FTYP box. Constructor sets default values
MP4::FTYP ftypBox;
if (sending3GP){
ftypBox.setMajorBrand("3gp6");
ftypBox.setCompatibleBrands("3gp6", 3);
}
header.write(ftypBox.asBox(), ftypBox.boxedSize());
//Start building the moov box. This is the metadata box for an mp4 file, and will contain all metadata.
MP4::MOOV moovBox;
//Keep track of the current index within the moovBox
unsigned int moovOffset = 0;
//Construct with duration of -1, as this is the default for fragmented
MP4::MVHD mvhdBox(-1);
//Then override it only when we are not sending a fragmented file
uint64_t fms;
if (!fragmented){
//calculating longest duration
uint64_t firstms = 0xFFFFFFFFFFFFFFull;
uint64_t lastms = 0;
for (std::set<unsigned long>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
lastms = std::max(lastms, (uint64_t)myMeta.tracks[*it].lastms);
firstms = std::min(firstms, (uint64_t)myMeta.tracks[*it].firstms);
}
mvhdBox.setDuration(lastms - firstms);
fms = firstms;
}
//Set the trackid for the first "empty" track within the file.
mvhdBox.setTrackID(selectedTracks.size() + 1);
moovBox.setContent(mvhdBox, moovOffset++);
for (std::set<unsigned long>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
DTSC::Track & thisTrack = myMeta.tracks[*it];
size_t partCount = thisTrack.parts.size();
uint64_t tDuration = thisTrack.lastms - thisTrack.firstms;
MP4::TRAK trakBox;
//Keep track of the current index within the moovBox
unsigned int trakOffset = 0;
MP4::TKHD tkhdBox(thisTrack, fragmented);
if (fragmented){
tkhdBox.setDuration(-1);
}
trakBox.setContent(tkhdBox, trakOffset++);
//Create an EDTS box, containing an ELST box with default values;
///\todo Figure out if this box is really needed for anything.
MP4::EDTS edtsBox;
MP4::ELST elstBox;
elstBox.setVersion(0);
elstBox.setFlags(0);
if (thisTrack.firstms != fms){
elstBox.setCount(2);
elstBox.setSegmentDuration(0, fragmented ? -1 : thisTrack.firstms - fms);
elstBox.setMediaTime(0, 0xFFFFFFFFull);
elstBox.setMediaRateInteger(0, 0);
elstBox.setMediaRateFraction(0, 0);
elstBox.setSegmentDuration(1, tDuration);
elstBox.setMediaTime(1, 0);
elstBox.setMediaRateInteger(1, 1);
elstBox.setMediaRateFraction(1, 0);
}else{
elstBox.setCount(1);
elstBox.setSegmentDuration(0, tDuration);
elstBox.setMediaTime(0, 0);
elstBox.setMediaRateInteger(0, 1);
elstBox.setMediaRateFraction(0, 0);
}
edtsBox.setContent(elstBox, 0);
trakBox.setContent(edtsBox, trakOffset++);
MP4::MDIA mdiaBox;
size_t mdiaOffset = 0;
//Add the mandatory MDHD and HDLR boxes to the MDIA
MP4::MDHD mdhdBox(tDuration);
if (fragmented){
mdhdBox.setDuration(-1);
}
mdhdBox.setLanguage(thisTrack.lang);
mdiaBox.setContent(mdhdBox, mdiaOffset++);
MP4::HDLR hdlrBox(thisTrack.type, thisTrack.getIdentifier());
mdiaBox.setContent(hdlrBox, mdiaOffset++);
MP4::MINF minfBox;
size_t minfOffset = 0;
MP4::STBL stblBox;
unsigned int stblOffset = 0;
//Add a track-type specific box to the MINF box
if (thisTrack.type == "video"){
MP4::VMHD vmhdBox;
vmhdBox.setFlags(1);
minfBox.setContent(vmhdBox, minfOffset++);
}else if (thisTrack.type == "audio"){
MP4::SMHD smhdBox;
minfBox.setContent(smhdBox, minfOffset++);
}else{
//create nmhd box
MP4::NMHD nmhdBox;
minfBox.setContent(nmhdBox, minfOffset++);
}
//Add the mandatory DREF (dataReference) box
MP4::DINF dinfBox;
MP4::DREF drefBox;
dinfBox.setContent(drefBox, 0);
minfBox.setContent(dinfBox, minfOffset++);
//Add STSD box
MP4::STSD stsdBox(0);
if (thisTrack.type == "video"){
MP4::VisualSampleEntry sampleEntry(thisTrack);
stsdBox.setEntry(sampleEntry, 0);
}else if (thisTrack.type == "audio"){
MP4::AudioSampleEntry sampleEntry(thisTrack);
stsdBox.setEntry(sampleEntry, 0);
}else if (thisTrack.type == "meta"){
INFO_MSG("add subtitlesample\n");
MP4::TextSampleEntry sampleEntry(thisTrack);
MP4::FontTableBox ftab;
sampleEntry.setFontTableBox(ftab);
stsdBox.setEntry(sampleEntry, 0);
}
stblBox.setContent(stsdBox, stblOffset++);
//Add STTS Box
//note: STTS is empty when fragmented
MP4::STTS sttsBox(0);
//Add STSZ Box
//note: STSZ is empty when fragmented
MP4::STSZ stszBox(0);
if (!fragmented){
MP4::CTTS cttsBox;
cttsBox.setVersion(0);
MP4::CTTSEntry tmpEntry;
tmpEntry.sampleCount = 0;
tmpEntry.sampleOffset = thisTrack.parts[0].getOffset();
std::deque<std::pair<size_t, size_t> > sttsCounter;
stszBox.setEntrySize(0, partCount - 1);//Speed up allocation
size_t totalEntries = 0;
for (size_t part = 0; part < partCount; ++part){
stats();
uint64_t partDur = thisTrack.parts[part].getDuration();
uint64_t partSize = thisTrack.parts[part].getSize();
uint64_t partOffset = thisTrack.parts[part].getOffset();
//Create a new entry with current duration if EITHER there is no entry yet, or this parts duration differs from the previous
if (!sttsCounter.size() || sttsCounter.rbegin()->second != partDur){
sttsCounter.push_back(std::pair<size_t,size_t>(0, partDur));
}
//Update the counter
sttsCounter.rbegin()->first++;
if(thisTrack.type == "meta"){
partSize += 2;
}
stszBox.setEntrySize(partSize, part);
size += partSize;
if (partOffset != tmpEntry.sampleOffset){
//If the offset of this and previous part differ, write current values and reset
cttsBox.setCTTSEntry(tmpEntry, totalEntries++);///\todo Again, rewrite for sanity. index FIRST, value SECOND
tmpEntry.sampleCount = 0;
tmpEntry.sampleOffset = partOffset;
}
tmpEntry.sampleCount++;
}
MP4::STTSEntry sttsEntry;
sttsBox.setSTTSEntry(sttsEntry, sttsCounter.size() - 1);
size_t sttsIdx = 0;
for (std::deque<std::pair<size_t, size_t> >::iterator it2 = sttsCounter.begin(); it2 != sttsCounter.end(); it2++){
sttsEntry.sampleCount = it2->first;
sttsEntry.sampleDelta = it2->second;
sttsBox.setSTTSEntry(sttsEntry, sttsIdx++);
}
if (totalEntries || tmpEntry.sampleOffset){
cttsBox.setCTTSEntry(tmpEntry, totalEntries++);
stblBox.setContent(cttsBox, stblOffset++);
}
}
stblBox.setContent(sttsBox, stblOffset++);
stblBox.setContent(stszBox, stblOffset++);
//Add STSS Box IF type is video and we are not fragmented
if (thisTrack.type == "video" && !fragmented){
MP4::STSS stssBox(0);
int tmpCount = 0;
for (int i = 0; i < thisTrack.keys.size(); i++){
stssBox.setSampleNumber(tmpCount + 1, i);///\todo PLEASE rewrite this for sanity.... SHOULD be: index FIRST, value SECOND
tmpCount += thisTrack.keys[i].getParts();
}
stblBox.setContent(stssBox, stblOffset++);
}
//Add STSC Box
//note: STSC is empty when fragmented
MP4::STSC stscBox(0);
if (!fragmented){
MP4::STSCEntry stscEntry(1,1,1);
stscBox.setSTSCEntry(stscEntry, 0);
}
stblBox.setContent(stscBox, stblOffset++);
//Create STCO Box (either stco or co64)
//note: 64bit boxes will never be used in fragmented
//note: Inserting empty values on purpose here, will be fixed later.
if (useLargeBoxes){
MP4::CO64 CO64Box;
CO64Box.setChunkOffset(0, partCount - 1);
stblBox.setContent(CO64Box, stblOffset++);
}else{
MP4::STCO stcoBox(0);
if (fragmented){
stcoBox.setEntryCount(0);
}else{
stcoBox.setChunkOffset(0, partCount - 1);
}
stblBox.setContent(stcoBox, stblOffset++);
}
minfBox.setContent(stblBox, minfOffset++);
mdiaBox.setContent(minfBox, mdiaOffset++);
trakBox.setContent(mdiaBox, 2);
moovBox.setContent(trakBox, moovOffset++);
}
if (fragmented){
MP4::MVEX mvexBox;
unsigned int curBox = 0;
for (std::set<long unsigned int>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
MP4::TREX trexBox(*it);
mvexBox.setContent(trexBox, curBox++);
}
moovBox.setContent(mvexBox, moovOffset++);
}else{//if we are making a non fragmented MP4 and there are parts
//initial offset length ftyp, length moov + 8
uint64_t dataOffset = ftypBox.boxedSize() + moovBox.boxedSize() + 8;
std::map <size_t, MP4::STCO> checkStcoBoxes;
std::map <size_t, MP4::CO64> checkCO64Boxes;
std::deque<MP4::TRAK> trak = moovBox.getChildren<MP4::TRAK>();
for (std::deque<MP4::TRAK>::iterator trakIt = trak.begin(); trakIt != trak.end(); trakIt++){
MP4::TKHD tkhdBox = trakIt->getChild<MP4::TKHD>();
MP4::STBL stblBox = trakIt->getChild<MP4::MDIA>().getChild<MP4::MINF>().getChild<MP4::STBL>();
if (useLargeBoxes){
checkCO64Boxes.insert(std::pair<size_t, MP4::CO64>(tkhdBox.getTrackID(), stblBox.getChild<MP4::CO64>()));
}else{
checkStcoBoxes.insert(std::pair<size_t, MP4::STCO>(tkhdBox.getTrackID(), stblBox.getChild<MP4::STCO>()));
}
}
//inserting right values in the STCO box header
//total = 0;
//Keep track of the current size of the data within the mdat
uint64_t dataSize = 0;
//Current values are actual byte offset without header-sized offset
std::set <keyPart> sortSet;//filling sortset for interleaving parts
for (std::set<long unsigned int>::iterator subIt = selectedTracks.begin(); subIt != selectedTracks.end(); subIt++){
DTSC::Track & thisTrack = myMeta.tracks[*subIt];
keyPart temp;
temp.trackID = *subIt;
temp.time = thisTrack.firstms;//timeplace of frame
temp.index = 0;
temp.size = thisTrack.parts[0].getDuration();
HIGH_MSG("Header sortSet: tid %lu time %lu", temp.trackID, temp.time);
sortSet.insert(temp);
}
while (!sortSet.empty()){
stats();
keyPart temp = *sortSet.begin();
sortSet.erase(sortSet.begin());
DTSC::Track & thisTrack = myMeta.tracks[temp.trackID];
//setting the right STCO size in the STCO box
if (useLargeBoxes){//Re-using the previously defined boolean for speedup
checkCO64Boxes[temp.trackID].setChunkOffset(dataOffset + dataSize, temp.index);
}else{
checkStcoBoxes[temp.trackID].setChunkOffset(dataOffset + dataSize, temp.index);
}
dataSize += thisTrack.parts[temp.index].getSize();
if(thisTrack.type == "meta"){
dataSize += 2;
}
//add next keyPart to sortSet
if (temp.index + 1< thisTrack.parts.size()){//Only create new element, when there are new elements to be added
temp.time += thisTrack.parts[temp.index].getDuration();
++temp.index;
temp.size = thisTrack.parts[temp.index].getSize();
sortSet.insert(temp);
}
}
///\todo Update this thing for boxes >4G?
mdatSize = dataSize + 8;//+8 for mp4 header
}
header << std::string(moovBox.asBox(), moovBox.boxedSize());
if (!fragmented){//if we are making a non fragmented MP4 and there are parts
char mdatHeader[8] ={0x00,0x00,0x00,0x00,'m','d','a','t'};
if (mdatSize < 0xFFFFFFFF){
Bit::htobl(mdatHeader, mdatSize);
}
header.write(mdatHeader, 8);
}else{
//this is a dirty fix to prevent the code from adding 0xDE to the end of the header
header << (char)(0);
}
size += header.str().size();
MEDIUM_MSG("Header %llu, file: %llu", header.str().size(), size);
if (fragmented){
realBaseOffset = header.str().size();
}
return header.str();
}
/// Calculate a seekPoint, based on byteStart, metadata, tracks and headerSize.
/// The seekPoint will be set to the timestamp of the first packet to send.
void OutProgressiveMP4::findSeekPoint(uint64_t byteStart, uint64_t & seekPoint, uint64_t headerSize){
seekPoint = 0;
//if we're starting in the header, seekPoint is always zero.
if (byteStart <= headerSize){
return;
}
//okay, we're past the header. Substract the headersize from the starting postion.
byteStart -= headerSize;
//forward through the file by headers, until we reach the point where we need to be
while (!sortSet.empty()){
//find the next part and erase it
keyPart temp = *sortSet.begin();
DTSC::Track & thisTrack = myMeta.tracks[temp.trackID];
uint64_t partSize = thisTrack.parts[temp.index].getSize();
//add 2 bytes in front of the subtitle that contains the length of the subtitle.
if(myMeta.tracks[temp.trackID].codec == "subtitle"){
partSize += 2;
}
//record where we are
seekPoint = temp.time;
//substract the size of this fragment from byteStart
//if that put us past the point where we wanted to be, return right now
if (partSize > byteStart){
INFO_MSG("We're starting at time %" PRIu64 ", skipping %" PRIu64 " bytes", seekPoint, partSize - byteStart);
return;
}
byteStart -= partSize;
//otherwise, set currPos to where we are now and continue
currPos += partSize;
if (temp.index + 1 < myMeta.tracks[temp.trackID].parts.size()){//only insert when there are parts left
temp.time += thisTrack.parts[temp.index].getDuration();
++temp.index;
temp.size = thisTrack.parts[temp.index].getSize();
sortSet.insert(temp);
}
//Remove just-parsed element
sortSet.erase(sortSet.begin());
//wash, rinse, repeat
}
//If we're here, we're in the last fragment.
//That's technically legal, of course.
}
void OutProgressiveMP4::sendFragmentHeader(){
uint64_t mdatSize = 8;
MP4::MOOF moofBox;
MP4::MFHD mfhdBox;
mfhdBox.setSequenceNumber(fragSeqNum++);
moofBox.setContent(mfhdBox, 0);
unsigned int moofIndex = 1;
sortSet.clear();
//sort all parts here
std::set <keyPart> trunOrder;
//set with trackID, relative data offset, time and size
for (std::map<size_t, fragSet>::iterator it = currentPartSet.begin(); it != currentPartSet.end(); it++){
uint64_t timeStamp = it->second.firstTime;
DTSC::Track & thisTrack = myMeta.tracks[it->first];
for (uint32_t i = it->second.firstPart; i <= it->second.lastPart; i++){
keyPart temp;
temp.trackID = it->first;
temp.time = timeStamp;
temp.index = i;
temp.size = thisTrack.parts[temp.index].getSize();
timeStamp += thisTrack.parts[temp.index].getDuration();
HIGH_MSG("Fragment sortSet: tid %lu time %lu", temp.trackID, temp.time);
trunOrder.insert(temp);
}
}
//now all the parts have been sorted, we make a relative ByteOffset
uint64_t relativeOffset = 0;
for (std::set<keyPart>::iterator it = trunOrder.begin(); it != trunOrder.end(); it++){
DTSC::Track & thisTrack = myMeta.tracks[it->trackID];
//We have to make a copy, because altering the element inside the set would invalidate the iterators
keyPart temp = *it;
temp.byteOffset = relativeOffset;
relativeOffset += temp.size;
DONTEVEN_MSG("Anticipating tid: %lu size: %lu", it->trackID, temp.size);
sortSet.insert(temp);
}
trunOrder.clear();//erase the trunOrder set, to keep memory usage down
bool firstSample = true;
bool hasAudio = false;
//Fun fact! Firefox cares about the ordering here.
//It doesn't care about the order or track IDs in the header.
//But - the first TRAF must be a video TRAF, if video is present.
std::deque<std::map<size_t, fragSet>::iterator> sortedTracks;
for (std::map<size_t, fragSet>::iterator it = currentPartSet.begin(); it != currentPartSet.end(); it++) {
if (myMeta.tracks[it->first].type == "video"){
sortedTracks.push_front(it);
}else{
if (myMeta.tracks[it->first].type == "audio"){
hasAudio = true;
}
sortedTracks.push_back(it);
}
}
for (std::deque<std::map<size_t, fragSet>::iterator>::iterator ti = sortedTracks.begin(); ti != sortedTracks.end(); ++ti) {
std::map<size_t, fragSet>::iterator & it = *ti;
size_t tid = it->first;
DTSC::Track & thisTrack = myMeta.tracks[tid];
MP4::TRAF trafBox;
MP4::TFHD tfhdBox;
tfhdBox.setFlags(MP4::tfhdBaseOffset | MP4::tfhdSampleDura | MP4::tfhdSampleSize | MP4::tfhdSampleFlag);
tfhdBox.setTrackID(tid);
tfhdBox.setBaseDataOffset(realBaseOffset - 1); //Offset of current moof box, we use currPos for this. Not sure why we need the -1, but this gives the right offset
tfhdBox.setDefaultSampleDuration(thisTrack.parts[it->second.firstPart].getDuration());
tfhdBox.setDefaultSampleSize(thisTrack.parts[it->second.firstPart].getSize());
if (tid == vidTrack){
tfhdBox.setDefaultSampleFlags(MP4::noIPicture | MP4::noKeySample);
}else{
tfhdBox.setDefaultSampleFlags(MP4::isIPicture | MP4::isKeySample);
}
trafBox.setContent(tfhdBox, 0);
unsigned int trafOffset = 1;
for (std::set<keyPart>::iterator trunIt = sortSet.begin(); trunIt != sortSet.end(); trunIt++){
if (trunIt->trackID == tid){
uint64_t partOffset = thisTrack.parts[trunIt->index].getOffset();
uint64_t partSize = thisTrack.parts[trunIt->index].getSize();
uint64_t partDur = thisTrack.parts[trunIt->index].getDuration();
MP4::TRUN trunBox;
trunBox.setFlags(MP4::trundataOffset | MP4::trunfirstSampleFlags | MP4::trunsampleSize | MP4::trunsampleDuration | (partOffset ? MP4::trunsampleOffsets : 0));
//The value set here, will be updated afterwards to the correct value
trunBox.setDataOffset(trunIt->byteOffset);
trunBox.setFirstSampleFlags(MP4::isIPicture | (firstSample ? MP4::isKeySample : MP4::noKeySample));
firstSample = false;
mdatSize += partSize;
MP4::trunSampleInformation sampleInfo;
sampleInfo.sampleSize = partSize;
sampleInfo.sampleDuration = partDur;
if (partOffset){
sampleInfo.sampleOffset = partOffset;
}
trunBox.setSampleInformation(sampleInfo, 0);
trafBox.setContent(trunBox, trafOffset++);
}
}
moofBox.setContent(trafBox, moofIndex);
moofIndex++;
}
//Oh god why do we do this.
if (chromeWorkaround && hasAudio && fragSeqNum == 1){
INFO_MSG("Activating Chrome MP4 compatibility workaround!");
MP4::TRAF trafBox;
MP4::TRUN trunBox;
trunBox.setFlags(MP4::trundataOffset | MP4::trunfirstSampleFlags | MP4::trunsampleSize | MP4::trunsampleDuration);
trunBox.setDataOffset(0);
trunBox.setFirstSampleFlags(MP4::isIPicture | MP4::noKeySample);
MP4::trunSampleInformation sampleInfo;
sampleInfo.sampleSize = 0;
sampleInfo.sampleDuration = -1;
trunBox.setSampleInformation(sampleInfo, 0);
trafBox.setContent(trunBox, 0);
moofBox.setContent(trafBox, moofIndex);
moofIndex++;
}
//Update the trun data offsets with their correct values
MP4::TRAF loopTrafBox;
MP4::TRUN fixTrunBox;
uint32_t moofCount = moofBox.getContentCount();
for (unsigned int i = 0; i < moofCount; i++){
if (moofBox.getContent(i).isType("traf")){
loopTrafBox = ((MP4::TRAF &)moofBox.getContent(i, true));
uint32_t trafCount = loopTrafBox.getContentCount();
for (unsigned int j = 0; j < trafCount; j++){
MP4::Box & tmpBox = loopTrafBox.getContent(j, true);
if (tmpBox.isType("trun")){
fixTrunBox = (MP4::TRUN &)tmpBox;
fixTrunBox.setDataOffset(fixTrunBox.getDataOffset() + moofBox.boxedSize() + 8);
}
}
}
}
realBaseOffset += (moofBox.boxedSize() + mdatSize);
myConn.SendNow(moofBox.asBox(), moofBox.boxedSize());
char mdatHeader[8] ={0x00,0x00,0x00,0x00,'m','d','a','t'};
Bit::htobl(mdatHeader, mdatSize);
myConn.SendNow(mdatHeader, 8);
}
void OutProgressiveMP4::onHTTP(){
if(H.method == "OPTIONS" || H.method == "HEAD"){
H.Clean();
H.setCORSHeaders();
H.SetHeader("Content-Type", "video/MP4");
H.SetHeader("Accept-Ranges", "bytes, parsec");
H.SendResponse("200", "OK", myConn);
return;
}
chromeWorkaround = (H.GetHeader("User-Agent").find("Chrome") != std::string::npos && H.GetHeader("User-Agent").find("Edge") == std::string::npos && H.GetHeader("User-Agent").find("OPR/") == std::string::npos);
/*LTS-START*/
//allow setting of max lead time through buffer variable.
//max lead time is set in MS, but the variable is in integer seconds for simplicity.
if (H.GetVar("buffer") != ""){
maxSkipAhead = JSON::Value(H.GetVar("buffer")).asInt() * 1000;
}
//allow setting of play back rate through buffer variable.
//play back rate is set in MS per second, but the variable is a simple multiplier.
if (H.GetVar("rate") != ""){
long long int multiplier = JSON::Value(H.GetVar("rate")).asInt();
if (multiplier){
realTime = 1000 / multiplier;
}else{
realTime = 0;
}
}
if (H.GetHeader("X-Mist-Rate") != ""){
long long int multiplier = JSON::Value(H.GetHeader("X-Mist-Rate")).asInt();
if (multiplier){
realTime = 1000 / multiplier;
}else{
realTime = 0;
}
}
/*LTS-END*/
//Make sure we start receiving data after this function
///\todo Should this happen here?
parseData = true;
wantRequest = false;
sentHeader = false;
//Check if the url contains .3gp --> if yes, we will send a 3gp header
sending3GP = (H.url.find(".3gp") != std::string::npos);
fileSize = 0;
uint64_t headerSize = mp4HeaderSize(fileSize, myMeta.live);
seekPoint = 0;
if (myMeta.live){
//for live we use fragmented mode
fragSeqNum = 0;
partListSent = 0;
partListLength = 0;
}
byteStart = 0;
byteEnd = fileSize - 1;
char rangeType = ' ';
currPos = 0;
sortSet.clear();
for (std::set<long unsigned int>::iterator subIt = selectedTracks.begin(); subIt != selectedTracks.end(); subIt++){
DTSC::Track & thisTrack = myMeta.tracks[*subIt];
keyPart temp;
temp.trackID = *subIt;
temp.time = thisTrack.firstms;//timeplace of frame
temp.index = 0;
temp.size = thisTrack.parts[temp.index].getSize();
sortSet.insert(temp);
}
if (!myMeta.live){
if (H.GetHeader("Range") != ""){
if (parseRange(byteStart, byteEnd)){
findSeekPoint(byteStart, seekPoint, headerSize);
}
rangeType = H.GetHeader("Range")[0];
}
}
H.Clean(); //make sure no parts of old requests are left in any buffers
H.setCORSHeaders();
H.SetHeader("Content-Type", "video/MP4"); //Send the correct content-type for MP4 files
if (myMeta.vod){
H.SetHeader("Accept-Ranges", "bytes, parsec");
}
if (rangeType != ' '){
if (!byteEnd){
if (rangeType == 'p'){
H.SetBody("Starsystem not in communications range");
H.SendResponse("416", "Starsystem not in communications range", myConn);
return;
}else{
H.SetBody("Requested Range Not Satisfiable");
H.SendResponse("416", "Requested Range Not Satisfiable", myConn);
return;
}
}else{
std::stringstream rangeReply;
rangeReply << "bytes " << byteStart << "-" << byteEnd << "/" << fileSize;
H.SetHeader("Content-Length", byteEnd - byteStart + 1);
H.SetHeader("Content-Range", rangeReply.str());
/// \todo Switch to chunked?
H.SendResponse("206", "Partial content", myConn);
//H.StartResponse("206", "Partial content", HTTP_R, conn);
}
}else{
if (myMeta.vod){
H.SetHeader("Content-Length", byteEnd - byteStart + 1);
}
/// \todo Switch to chunked?
H.SendResponse("200", "OK", myConn);
//HTTP_S.StartResponse(HTTP_R, conn);
}
leftOver = byteEnd - byteStart + 1;//add one byte, because range "0-0" = 1 byte of data
if (byteStart < headerSize){
std::string headerData = DTSCMeta2MP4Header(fileSize, myMeta.live);
myConn.SendNow(headerData.data() + byteStart, std::min(headerSize, byteEnd) - byteStart); //send MP4 header
leftOver -= std::min(headerSize, byteEnd) - byteStart;
}
currPos += headerSize;//we're now guaranteed to be past the header point, no matter what
}
///Builds up a datastructure that allows for access in the fragment send header function
///using the fragment number **FOR THIS USER, NOT ACTUAL FRAGMENT NUMBER, HAS NOTHING TO DO WITH ACTUAL FRAGMENTS EVEN**
///We take the corresponding keyframe and interframes of the main video track and take concurrent frames from its secondary (audio) tracks
///\todo See if we can use something more elegant than a member variable...
void OutProgressiveMP4::buildFragment(){
if (!needsLookAhead){
needsLookAhead = 1000;
currentPartSet.clear();
return;
}
DTSC::Key & currKey = myMeta.tracks[vidTrack].getKey(getKeyForTime(vidTrack, thisPacket.getTime()));
uint64_t startms = thisPacket.getTime();
uint64_t endms = startms + needsLookAhead;
bool missingSome = true;
while (missingSome){
missingSome = false;
currentPartSet.clear();
for (std::set<long unsigned int>::iterator it = selectedTracks.begin(); it != selectedTracks.end(); it++){
DTSC::Track & thisTrack = myMeta.tracks[*it];
fragSet thisRange;
thisRange.firstPart = 0;
thisRange.firstTime = thisTrack.keys.begin()->getTime();
unsigned long long int prevParts = 0;
for (std::deque<DTSC::Key>::iterator it2 = thisTrack.keys.begin(); it2 != thisTrack.keys.end(); it2++){
if (it2->getTime() > startms){
break;
}
thisRange.firstPart += prevParts;
prevParts = it2->getParts();
thisRange.firstTime = it2->getTime();
}
thisRange.lastPart = thisRange.firstPart;
thisRange.lastTime = thisRange.firstTime;
uint64_t curMS = thisRange.firstTime;
uint64_t nextMS = thisRange.firstTime;
bool first = true;
size_t maxParts = thisTrack.parts.size();
for (size_t i = thisRange.firstPart; i < maxParts; i++){
if (first && curMS >= startms){
thisRange.firstPart = i;
thisRange.firstTime = curMS;
first = false;
}
uint32_t pDur = thisTrack.parts[i].getDuration();
nextMS = curMS + pDur;
//Make sure we always look ahead at least a single frame
if (pDur > needsLookAhead){
needsLookAhead = pDur;
INFO_MSG("Slow frame! Increasing lookAhead to %ums", needsLookAhead);
}
thisRange.lastPart = i;
thisRange.lastTime = curMS;
if (nextMS >= endms){break;}
curMS = nextMS;
}
if (first){
endms = thisTrack.lastms;
if (endms > startms && needsLookAhead < endms - startms){
needsLookAhead = endms - startms;
INFO_MSG("False start! Increasing lookAhead to %ums", needsLookAhead);
missingSome = true;
}
break;
}
currentPartSet[*it] = thisRange;
}
}
}
void OutProgressiveMP4::sendNext(){
static bool perfect = true;
//Obtain a pointer to the data of this packet
char * dataPointer = 0;
unsigned int len = 0;
thisPacket.getString("data", dataPointer, len);
std::string subtitle;
if (myMeta.live){
//if header needed
if (!partListLength || partListSent >= partListLength){
if (fragSeqNum > 10){
if (liveSeek()){return;}
}
//building set first
buildFragment();//map with metadata for keyframe
sendFragmentHeader();
partListSent = 0;
partListLength = 0;
for (std::map<size_t, fragSet>::iterator it = currentPartSet.begin(); it != currentPartSet.end(); it++){
partListLength += it->second.lastPart - it->second.firstPart + 1;
}
}
//generate content in mdat, meaning: send right parts
DONTEVEN_MSG("Sending tid: %ld size: %u", thisPacket.getTrackId() , len);
myConn.SendNow(dataPointer, len);
partListSent++;
}
keyPart thisPart = *sortSet.begin();
if ((unsigned long)thisPacket.getTrackId() != thisPart.trackID || thisPacket.getTime() != thisPart.time || len != thisPart.size){
if (thisPacket.getTime() > sortSet.begin()->time || thisPacket.getTrackId() > sortSet.begin()->trackID){
if (perfect){
WARN_MSG("Warning: input is inconsistent. Expected %lu:%lu but got %ld:%llu - cancelling playback", thisPart.trackID, thisPart.time, thisPacket.getTrackId(), thisPacket.getTime());
perfect = false;
myConn.close();
}
}else{
WARN_MSG("Did not receive expected %lu:%lu (%lub) but got %ld:%llu (%ub) - throwing it away", thisPart.trackID, thisPart.time, thisPart.size, thisPacket.getTrackId(), thisPacket.getTime(), len);
}
return;
}
//The remainder of this function handles non-live situations
if (myMeta.live){
sortSet.erase(sortSet.begin());
return;
}
//prepend subtitle text with 2 bytes datalength
if(myMeta.tracks[thisPacket.getTrackId()].codec == "subtitle"){
char pre[2];
Bit::htobs(pre,len);
subtitle.assign(pre,2);
subtitle.append(dataPointer, len);
dataPointer = (char*)subtitle.c_str();
len+=2;
}
if (currPos >= byteStart){
myConn.SendNow(dataPointer, std::min(leftOver, (int64_t)len));
leftOver -= len;
}else{
if (currPos + (long long)len > byteStart){
myConn.SendNow(dataPointer + (byteStart - currPos), std::min(leftOver, (int64_t)(len - (byteStart - currPos))));
leftOver -= len - (byteStart - currPos);
}
}
//keep track of where we are
if (!sortSet.empty()){
keyPart temp = *sortSet.begin();
sortSet.erase(sortSet.begin());
DTSC::Track & thisTrack = myMeta.tracks[temp.trackID];
currPos += thisTrack.parts[temp.index].getSize();
if (temp.index + 1 < thisTrack.parts.size()){//only insert when there are parts left
temp.time += thisTrack.parts[temp.index].getDuration();
++temp.index;
temp.size = thisTrack.parts[temp.index].getSize();
sortSet.insert(temp);
}
}
if (leftOver < 1){
//stop playback, wait for new request
stop();
wantRequest = true;
}
}
void OutProgressiveMP4::sendHeader(){
if (myMeta.live){
vidTrack = getMainSelectedTrack();
bool reSeek = false;
DTSC::Track & Trk = myMeta.tracks[vidTrack];
for (int i = 0; i < Trk.parts.size(); i++){
uint32_t pDur = Trk.parts[i].getDuration();
//Make sure we always look ahead at least a single frame
if (pDur > needsLookAhead){
needsLookAhead = pDur;
reSeek = true;
}
}
if (reSeek){
INFO_MSG("Increased initial lookAhead of %ums", needsLookAhead);
initialSeek();
}
}else{
seek(seekPoint);
}
sentHeader = true;
}
}