a
    PSicXB                     @   s   d dl Z d dlZd dlZd dlmZ d dlmZmZmZm	Z	m
Z
mZmZmZmZ d dlZd dlmZmZmZmZ ddlmZ edZejfeeee
edd	d
ZdejeeeejdddZG dd dZeedddZG dd dZdS )    NFraction)	AnyDictListOptionalCallableUnionTupleTypeVarcast)_probe_video_from_file_read_video_from_file
read_videoread_video_timestamps   )tqdmT)ptstimebase_fromtimebase_to
round_funcreturnc                 C   s   t | d| | }||S )zconvert pts between different time bases
    Args:
        pts: presentation timestamp, float
        timebase_from: original timebase. Fraction
        timebase_to: new timebase. Fraction
        round_func: rounding function.
    r   r   )r   r   r   r   new_pts r   \/var/www/html/django/DPS/env/lib/python3.9/site-packages/torchvision/datasets/video_utils.pypts_convert   s    r   )tensorsizestepdilationr   c                 C   s   |   dkrtd|    | d}|  }|| || f}|||d  d  | d |f}|d dk rtd|f}t| ||S )a   
    similar to tensor.unfold, but with the dilation
    and specialized for 1d tensors

    Returns all consecutive windows of `size` elements, with
    `step` between windows. The distance between each element
    in a window is given by `dilation`.
    r   z*tensor should have 1 dimension instead of r   )dim
ValueErrorstridenumeltorch
as_strided)r   r   r   r    Zo_strider$   Z
new_stridenew_sizer   r   r   unfold    s    	
 r(   c                   @   sR   e Zd ZdZee ddddZedddZee	ee e
e f d	d
dZdS )_VideoTimestampsDatasetz
    Dataset used to parallelize the reading of the timestamps
    of a list of videos, given their paths in the filesystem.

    Used in VideoClips and defined at top level so it can be
    pickled when forking.
    N)video_pathsr   c                 C   s
   || _ d S Nr*   )selfr*   r   r   r   __init__=   s    z _VideoTimestampsDataset.__init__r   c                 C   s
   t | jS r+   lenr*   r-   r   r   r   __len__@   s    z_VideoTimestampsDataset.__len__idxr   c                 C   s   t | j| S r+   )r   r*   )r-   r5   r   r   r   __getitem__C   s    z#_VideoTimestampsDataset.__getitem__)__name__
__module____qualname____doc__r   strr.   intr3   r
   r   floatr6   r   r   r   r   r)   4   s   r)   )xr   c                 C   s   | S )zH
    Dummy collate function to be used with _VideoTimestampsDataset
    r   )r>   r   r   r   _collate_fnG   s    r?   c                   @   s  e Zd ZdZd.ee eeee eeee	f  eeeeeeeedddd	Z
dd
ddZeee	f ddddZeeee	f d
ddZee d dddZed/ejeeeee eejeee ejf f dddZd0eeee ddddZed
ddZed
ddZed
dd Zeeeef d!d"d#Zeeeeeeejf d$d%d&Zeeejejeee	f ef d!d'd(Zeee	f d
d)d*Zeee	f dd+d,d-ZdS )1
VideoClipsaC  
    Given a list of video files, computes all consecutive subvideos of size
    `clip_length_in_frames`, where the distance between each subvideo in the
    same video is defined by `frames_between_clips`.
    If `frame_rate` is specified, it will also resample all the videos to have
    the same frame rate, and the clips will refer to this frame rate.

    Creating this instance the first time is time-consuming, as it needs to
    decode all the videos in `video_paths`. It is recommended that you
    cache the results after instantiation of the class.

    Recreating the clips for different clip lengths is fast, and can be done
    with the `compute_clips` method.

    Args:
        video_paths (List[str]): paths to the video files
        clip_length_in_frames (int): size of a clip in number of frames
        frames_between_clips (int): step (in frames) between each clip
        frame_rate (int, optional): if specified, it will resample the video
            so that it has `frame_rate`, and then the clips will be defined
            on the resampled video
        num_workers (int): how many subprocesses to use for data loading.
            0 means that the data will be loaded in the main process. (default: 0)
        output_format (str): The format of the output video tensors. Can be either "THWC" (default) or "TCHW".
       r   Nr   THWC)r*   clip_length_in_framesframes_between_clips
frame_rate_precomputed_metadatanum_workers_video_width_video_height_video_min_dimension_video_max_dimension_audio_samples_audio_channelsoutput_formatr   c                 C   s   || _ || _|| _|| _|	| _|
| _|| _|| _| | _	| j	dvrTt
d| d|d u rf|   n
| | | ||| d S )N)rB   TCHWz5output_format should be either 'THWC' or 'TCHW', got .)r*   rG   rH   rI   rJ   rK   rL   rM   upperrN   r"   _compute_frame_pts_init_from_metadatacompute_clips)r-   r*   rC   rD   rE   rF   rG   rH   rI   rJ   rK   rL   rM   rN   r   r   r   r.   i   s    



zVideoClips.__init__r/   c                    s   g | _ g | _dd l  jjjt| jd| jt	d}t
t|d^}|D ]H}|d tt| \}} fdd|D }| j | | j| qFW d    n1 s0    Y  d S )Nr   rA   )
batch_sizerG   
collate_fn)totalr   c                    s   g | ]} j | jd qS )dtype)	as_tensorlong).0cr%   r   r   
<listcomp>       z1VideoClips._compute_frame_pts.<locals>.<listcomp>)	video_pts	video_fpstorch.utils.datautilsdata
DataLoaderr)   r*   rG   r?   r   r1   updatelistzipextend)r-   dlpbarbatchclipsfpsr   r^   r   rR      s     
zVideoClips._compute_frame_pts)metadatar   c                 C   sV   |d | _ t| j t|d ks$J |d | _t| j t|d ksHJ |d | _d S )Nr*   ra   rb   )r*   r1   ra   rb   )r-   rp   r   r   r   rS      s
    

zVideoClips._init_from_metadatac                 C   s   | j | j| jd}|S )Nr*   ra   rb   rq   )r-   	_metadatar   r   r   rp      s
    zVideoClips.metadata)indicesr   c                    sz    fdd|D } fdd|D } fdd|D }|||d}t  | j j j| j j j j j j	 j
dS )Nc                    s   g | ]} j | qS r   r,   r\   ir2   r   r   r_      r`   z%VideoClips.subset.<locals>.<listcomp>c                    s   g | ]} j | qS r   )ra   rt   r2   r   r   r_      r`   c                    s   g | ]} j | qS r   )rb   rt   r2   r   r   r_      r`   rq   )rF   rG   rH   rI   rJ   rK   rL   rM   )type
num_framesr   rE   rG   rH   rI   rJ   rK   rL   rM   )r-   rs   r*   ra   rb   rp   r   r2   r   subset   s*    zVideoClips.subset)ra   rw   r   ro   rE   r   c           	      C   s   |d u rd}|d u r|}t | t||  }ttt|||}| | } t| ||}| sjt	
d t|tr|gt | }nt|||}||fS )Nr   zThere aren't enough frames in the current video to get a clip for the given clip length and frames between clips. The video (and potentially others) will be skipped.)r1   r=   r@   _resample_video_idxr<   mathfloorr(   r$   warningswarn
isinstanceslice)	ra   rw   r   ro   rE   Ztotal_framesZ_idxsrn   idxsr   r   r   compute_clips_for_video   s     
z"VideoClips.compute_clips_for_video)rw   r   rE   r   c           	      C   s   || _ || _|| _g | _g | _t| j| jD ]6\}}| |||||\}}| j	| | j	| q,t
dd | jD }|d | _dS )a  
        Compute all consecutive sequences of clips from video_pts.
        Always returns clips of size `num_frames`, meaning that the
        last few frames in a video can potentially be dropped.

        Args:
            num_frames (int): number of frames for the clip
            step (int): distance between two clips
            frame_rate (int, optional): The frame rate
        c                 S   s   g | ]}t |qS r   r1   r\   vr   r   r   r_      r`   z,VideoClips.compute_clips.<locals>.<listcomp>r   N)rw   r   rE   rn   resampling_idxsri   ra   rb   r   appendr%   rZ   cumsumtolistcumulative_sizes)	r-   rw   r   rE   ra   ro   rn   r   Zclip_lengthsr   r   r   rT      s    zVideoClips.compute_clipsc                 C   s   |   S r+   )	num_clipsr2   r   r   r   r3     s    zVideoClips.__len__c                 C   s
   t | jS r+   r0   r2   r   r   r   
num_videos  s    zVideoClips.num_videosc                 C   s
   | j d S )zJ
        Number of subclips that are available in the video list.
        )r   r2   r   r   r   r     s    zVideoClips.num_clipsr4   c                 C   s6   t | j|}|dkr|}n|| j|d   }||fS )zw
        Converts a flattened representation of the indices into a video_idx, clip_idx
        representation.
        r   r   )bisectbisect_rightr   )r-   r5   	video_idxclip_idxr   r   r   get_clip_location  s
    zVideoClips.get_clip_location)rw   original_fpsnew_fpsr   c                 C   sP   t || }| r(t|}td d |S tj| tjd| }| tj	}|S )NrX   )
r=   
is_integerr<   r   r%   arangefloat32r{   toint64)rw   r   r   r   r   r   r   r   ry     s    zVideoClips._resample_video_idxc                 C   st  ||   kr&td| d|    d| |\}}| j| }| j| | }ddlm} | }|dkr| jdkrxtd| j	dkrtd| j
dkrtd	| jdkrtd
| jdkrtd|dkr|d  }|d  }	t|||	\}
}}nt|}|j}d}tt|d  }tt|d  }d\}}tdd}t|jj|jj}|jrt|jj|jj}t|||tj}t|||tj}|j}t|| j| j	| j
| j||f|| j| j||f|d\}
}}d|i}|dur||d< | j dur&| j!| | }t"|t#j$r||d  }|
| }
| j |d< t%|
| j&ksLJ |
j' d| j& | j(dkrh|
)dddd}
|
|||fS )a7  
        Gets a subclip from a list of videos.

        Args:
            idx (int): index of the subclip. Must be between 0 and num_clips().

        Returns:
            video (Tensor)
            audio (Tensor)
            info (Dict)
            video_idx (int): index of the video in `video_paths`
        zIndex z out of range (z number of clips)r   )get_video_backendpyavz.pyav backend doesn't support _video_width != 0z/pyav backend doesn't support _video_height != 0z6pyav backend doesn't support _video_min_dimension != 0z6pyav backend doesn't support _video_max_dimension != 0z0pyav backend doesn't support _audio_samples != 0r   N)r   r   r   )
video_widthvideo_heightvideo_min_dimensionvideo_max_dimensionvideo_pts_rangevideo_timebaseaudio_samplesaudio_channelsaudio_pts_rangeaudio_timebaserb   	audio_fpsz x rO         )*r   
IndexErrorr   r*   rn   torchvisionr   rH   r"   rI   rJ   rK   rL   itemr   r   rb   r   r<   r   r   	numeratordenominator	has_audior   r   rz   r{   ceilaudio_sample_rater   rM   rE   r   r~   r%   Tensorr1   rw   shaperN   permute)r-   r5   r   r   
video_pathZclip_ptsr   backend	start_ptsend_ptsvideoaudioinfo_inforb   r   Zvideo_start_ptsZvideo_end_ptsZaudio_start_ptsZaudio_end_ptsr   r   _Zresampling_idxr   r   r   get_clip%  sz    








&zVideoClips.get_clipc                 C   sn   dd | j D }dd | j D }|r6t|}| }| j }||d< ||d< |d= |d= |d= d	|d
< |S )Nc                 S   s   g | ]}t |qS r   r   r   r   r   r   r_   ~  r`   z+VideoClips.__getstate__.<locals>.<listcomp>c                 S   s   g | ]}| tjqS r   )r   r%   r   )r\   r>   r   r   r   r_     r`   video_pts_sizesra   rn   r   r   r   _version)ra   r%   catnumpy__dict__copy)r-   r   ra   dr   r   r   __getstate__}  s    

zVideoClips.__getstate__)r   r   c                 C   sf   d|vr|| _ d S tj|d tjd}tj||d dd}|d= ||d< || _ | | j| j| j d S )Nr   ra   rX   r   r   )r!   )	r   r%   rZ   r   splitrT   rw   r   rE   )r-   r   ra   r   r   r   __setstate__  s    zVideoClips.__setstate__)rA   r   NNr   r   r   r   r   r   r   rB   )N)N) r7   r8   r9   r:   r   r;   r<   r   r   r   r.   rR   rS   propertyrp   rx   staticmethodr%   r   r
   r	   r   r   rT   r3   r   r   r   ry   r   r   r   r   r   r   r   r@   N   sd               %  (Xr@   )r   ) r   rz   r|   	fractionsr   typingr   r   r   r   r   r	   r
   r   r   r%   torchvision.ior   r   r   r   rd   r   r   r{   r<   r   r   r(   r)   r?   r@   r   r   r   r   <module>   s   ,