使用Java怎么样实现一个合并MP4视频功能

发布时间:2020-11-20 14:52:45 作者:Leah
来源:亿速云 阅读:731

今天就跟大家聊聊有关使用Java怎么样实现一个合并MP4视频功能,可能很多人都不太了解,为了让大家更加了解,小编给大家总结了以下内容,希望大家根据这篇文章可以有所收获。

局限性

只支持MP4文件
经过尝试对于一些MP4文件分割不了

依赖

<!-- mp4文件操作jar -->
<!-- https://mvnrepository.com/artifact/com.googlecode.mp4parser/isoparser -->
<dependency>
<groupId>com.googlecode.mp4parser</groupId>
<artifactId>isoparser</artifactId>
<version>1.1.22</version>
</dependency>

工具类

package com.example.demo;


import com.coremedia.iso.boxes.Container;
import com.googlecode.mp4parser.authoring.Movie;
import com.googlecode.mp4parser.authoring.Track;
import com.googlecode.mp4parser.authoring.builder.DefaultMp4Builder;
import com.googlecode.mp4parser.authoring.container.mp4.MovieCreator;
import com.googlecode.mp4parser.authoring.tracks.AppendTrack;
import com.googlecode.mp4parser.authoring.tracks.CroppedTrack;

import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
import java.nio.channels.FileChannel;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.LinkedList;
import java.util.List;

public class Mp4ParserUtils {
 /**
  * 合并视频
  *
  * @param videoList: 所有视频地址集合
  * @param mergeVideoFile: 目标文件
  * @return
  */
 public static String mergeVideo(List<String> videoList, File mergeVideoFile) {
  FileOutputStream fos = null;
  FileChannel fc = null;
  try {
   List<Movie> sourceMovies = new ArrayList<>();
   for (String video : videoList) {
    sourceMovies.add(MovieCreator.build(video));
   }

   List<Track> videoTracks = new LinkedList<>();
   List<Track> audioTracks = new LinkedList<>();

   for (Movie movie : sourceMovies) {
    for (Track track : movie.getTracks()) {
     if ("soun".equals(track.getHandler())) {
      audioTracks.add(track);
     }

     if ("vide".equals(track.getHandler())) {
      videoTracks.add(track);
     }
    }
   }

   Movie mergeMovie = new Movie();
   if (audioTracks.size() > 0) {
    mergeMovie.addTrack(new AppendTrack(audioTracks.toArray(new Track[audioTracks.size()])));
   }

   if (videoTracks.size() > 0) {
    mergeMovie.addTrack(new AppendTrack(videoTracks.toArray(new Track[videoTracks.size()])));
   }

   Container out = new DefaultMp4Builder().build(mergeMovie);
   fos = new FileOutputStream(mergeVideoFile);
   fc = fos.getChannel();
   out.writeContainer(fc);
   fc.close();
   fos.close();
   return mergeVideoFile.getAbsolutePath();
  } catch (Exception e) {
   e.printStackTrace();
  } finally {
   if (fc != null) {
    try {
     fc.close();
    } catch (IOException e) {
     e.printStackTrace();
    }
   }

   if (fos != null) {
    try {
     fos.close();
    } catch (IOException e) {
     e.printStackTrace();
    }
   }
  }

  return null;
 }

 /**
  * 剪切视频
  * @param srcVideoPath
  * @param dstVideoPath
  * @param times
  * @throws IOException
  */
 public static void cutVideo(String srcVideoPath, String dstVideoPath, double[] times) throws IOException {
  int dstVideoNumber = times.length / 2;
  String[] dstVideoPathes = new String[dstVideoNumber];
  for (int i = 0; i < dstVideoNumber; i++) {
   dstVideoPathes[i] = dstVideoPath + "cutOutput-" + i + ".mp4";
  }
  int timesCount = 0;

  for (int idst = 0; idst < dstVideoPathes.length; idst++) {
   //Movie movie = new MovieCreator().build(new RandomAccessFile("/home/sannies/suckerpunch-distantplanet_h2080p/suckerpunch-distantplanet_h2080p.mov", "r").getChannel());
   Movie movie = MovieCreator.build(srcVideoPath);

   List<Track> tracks = movie.getTracks();
   movie.setTracks(new LinkedList<Track>());
   // remove all tracks we will create new tracks from the old


   double startTime1 = times[timesCount];
   double endTime1 = times[timesCount + 1];
   timesCount = timesCount + 2;

   boolean timeCorrected = false;

   // Here we try to find a track that has sync samples. Since we can only start decoding
   // at such a sample we SHOULD make sure that the start of the new fragment is exactly
   // such a frame
   for (Track track : tracks) {
    if (track.getSyncSamples() != null && track.getSyncSamples().length > 0) {
     if (timeCorrected) {
      // This exception here could be a false positive in case we have multiple tracks
      // with sync samples at exactly the same positions. E.g. a single movie containing
      // multiple qualities of the same video (Microsoft Smooth Streaming file)

      throw new RuntimeException("The startTime has already been corrected by another track with SyncSample. Not Supported.");
     }
     startTime1 = correctTimeToSyncSample(track, startTime1, false);
     endTime1 = correctTimeToSyncSample(track, endTime1, true);

     timeCorrected = true;
    }
   }

   for (Track track : tracks) {
    long currentSample = 0;
    double currentTime = 0;
    double lastTime = -1;
    long startSample1 = -1;
    long endSample1 = -1;


    for (int i = 0; i < track.getSampleDurations().length; i++) {
     long delta = track.getSampleDurations()[i];


     if (currentTime > lastTime && currentTime <= startTime1) {
      // current sample is still before the new starttime
      startSample1 = currentSample;
     }
     if (currentTime > lastTime && currentTime <= endTime1) {
      // current sample is after the new start time and still before the new endtime
      endSample1 = currentSample;
     }

     lastTime = currentTime;
     currentTime += (double) delta / (double) track.getTrackMetaData().getTimescale();
     currentSample++;
    }
    //movie.addTrack(new AppendTrack(new ClippedTrack(track, startSample1, endSample1), new ClippedTrack(track, startSample2, endSample2)));
    movie.addTrack(new CroppedTrack(track, startSample1, endSample1));
   }
   long start1 = System.currentTimeMillis();
   Container out = new DefaultMp4Builder().build(movie);
   long start2 = System.currentTimeMillis();
   FileOutputStream fos = new FileOutputStream(String.format(dstVideoPathes[idst]));
   FileChannel fc = fos.getChannel();
   out.writeContainer(fc);

   fc.close();
   fos.close();
   long start3 = System.currentTimeMillis();

  }
 }

 private static double correctTimeToSyncSample(Track track, double cutHere, boolean next) {
  double[] timeOfSyncSamples = new double[track.getSyncSamples().length];
  long currentSample = 0;
  double currentTime = 0;
  for (int i = 0; i < track.getSampleDurations().length; i++) {
   long delta = track.getSampleDurations()[i];

   if (Arrays.binarySearch(track.getSyncSamples(), currentSample + 1) >= 0) {
    // samples always start with 1 but we start with zero therefore +1
    timeOfSyncSamples[Arrays.binarySearch(track.getSyncSamples(), currentSample + 1)] = currentTime;
   }
   currentTime += (double) delta / (double) track.getTrackMetaData().getTimescale();
   currentSample++;

  }
  double previous = 0;
  for (double timeOfSyncSample : timeOfSyncSamples) {
   if (timeOfSyncSample > cutHere) {
    if (next) {
     return timeOfSyncSample;
    } else {
     return previous;
    }
   }
   previous = timeOfSyncSample;
  }
  return timeOfSyncSamples[timeOfSyncSamples.length - 1];
 }

}

看完上述内容,你们对使用Java怎么样实现一个合并MP4视频功能有进一步的了解吗?如果还想了解更多知识或者相关内容,请关注亿速云行业资讯频道,感谢大家的支持。

推荐阅读:
  1. python opencv如何读mp4视频
  2. 如何使用JS基于对象实现合并功能

免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。

java ava

上一篇:修改harbor配置文件后重启docker的方法

下一篇:excel数据使用Datagrip怎么样实现导入

相关阅读

您好,登录后才能下订单哦!

密码登录
登录注册
其他方式登录
点击 登录注册 即表示同意《亿速云用户服务条款》