
    fThL              	          S r SSKJrJrJr  SSKrSSKrSSKJr  SSKJ	r	J
r
Jr  SSKJr  SSKJrJrJrJr  SS	KJr  SS
KJrJr  SSKJr  SSKJr  \R8                  " \5      rS+S\R>                  S\ S\!S\R>                  4S jjr" " S S\RF                  5      r$ " S S\RF                  5      r% " S S\RF                  5      r& " S S\RF                  5      r' " S S\RF                  5      r( " S S\RF                  5      r)\ " S S \5      5       r*\ " S! S"\*5      5       r+\" S#S$9 " S% S&\*5      5       r,\" S'S$9 " S( S)\*\5      5       r-/ S*Qr.g),zPyTorch ConvNext model.    )OptionalTupleUnionN)nn)BCEWithLogitsLossCrossEntropyLossMSELoss   )ACT2FN)BackboneOutputBaseModelOutputWithNoAttention(BaseModelOutputWithPoolingAndNoAttention$ImageClassifierOutputWithNoAttention)PreTrainedModel)auto_docstringlogging)BackboneMixin   )ConvNextConfiginput	drop_probtrainingreturnc                    US:X  d  U(       d  U $ SU-
  nU R                   S   4SU R                  S-
  -  -   nU[        R                  " X@R                  U R
                  S9-   nUR                  5         U R                  U5      U-  nU$ )a*  
Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks).

Comment by Ross Wightman: This is the same as the DropConnect impl I created for EfficientNet, etc networks,
however, the original name is misleading as 'Drop Connect' is a different form of dropout in a separate paper...
See discussion: https://github.com/tensorflow/tpu/issues/494#issuecomment-532968956 ... I've opted for changing the
layer and argument names to 'drop path' rather than mix DropConnect as a layer name and use 'survival rate' as the
argument.
        r   r   )r   )dtypedevice)shapendimtorchrandr   r   floor_div)r   r   r   	keep_probr   random_tensoroutputs          f/var/www/auris/envauris/lib/python3.13/site-packages/transformers/models/convnext/modeling_convnext.py	drop_pathr(   )   s     CxII[[^

Q 77E

5ELL YYMYYy!M1FM    c                      ^  \ rS rSrSrSS\\   SS4U 4S jjjrS\R                  S\R                  4S jr
S\4S	 jrS
rU =r$ )ConvNextDropPath>   zXDrop paths (Stochastic Depth) per sample (when applied in main path of residual blocks).Nr   r   c                 .   > [         TU ]  5         Xl        g N)super__init__r   )selfr   	__class__s     r'   r0   ConvNextDropPath.__init__A   s    "r)   hidden_statesc                 B    [        XR                  U R                  5      $ r.   )r(   r   r   r1   r4   s     r'   forwardConvNextDropPath.forwardE   s    FFr)   c                 8    SR                  U R                  5      $ )Nzp={})formatr   )r1   s    r'   
extra_reprConvNextDropPath.extra_reprH   s    }}T^^,,r)   )r   r.   )__name__
__module____qualname____firstlineno____doc__r   floatr0   r    Tensorr7   strr;   __static_attributes____classcell__r2   s   @r'   r+   r+   >   sQ    b#(5/ #T # #GU\\ Gell G-C - -r)   r+   c                   j   ^  \ rS rSrSrSU 4S jjrS\R                  S\R                  4S jrSr	U =r
$ )	ConvNextLayerNormL   a5  LayerNorm that supports two data formats: channels_last (default) or channels_first.
The ordering of the dimensions in the inputs. channels_last corresponds to inputs with shape (batch_size, height,
width, channels) while channels_first corresponds to inputs with shape (batch_size, channels, height, width).
c                 V  > [         TU ]  5         [        R                  " [        R
                  " U5      5      U l        [        R                  " [        R                  " U5      5      U l        X l	        X0l
        U R                  S;  a  [        SU R                   35      eU4U l        g )N)channels_lastchannels_firstzUnsupported data format: )r/   r0   r   	Parameterr    onesweightzerosbiasepsdata_formatNotImplementedErrornormalized_shape)r1   rV   rS   rT   r2   s       r'   r0   ConvNextLayerNorm.__init__R   s    ll5::.>#?@LL-=!>?	&#FF%(A$BRBRAS&TUU!1 3r)   xr   c                 P   U R                   S:X  aV  [        R                  R                  R	                  XR
                  U R                  U R                  U R                  5      nU$ U R                   S:X  a  UR                  nUR                  5       nUR                  SSS9nX-
  R                  S5      R                  SSS9nX-
  [        R                  " X@R                  -   5      -  nUR                  US9nU R                  S S 2S S 4   U-  U R                  S S 2S S 4   -   nU$ )NrL   rM   r   T)keepdim   )r   )rT   r    r   
functional
layer_normrV   rP   rR   rS   r   rB   meanpowsqrtto)r1   rX   input_dtypeuss        r'   r7   ConvNextLayerNorm.forward\   s   .##..q2G2GVZV_V_aeaiaijA  !11''K	Aq$'AA##At#4A%**Q\22A;'AAtTM*Q.1dD=1IIAr)   )rR   rT   rS   rV   rP   )ư>rL   )r=   r>   r?   r@   rA   r0   r    rC   r7   rE   rF   rG   s   @r'   rI   rI   L   s-    
4 %,,  r)   rI   c                   f   ^  \ rS rSrSrU 4S jrS\R                  S\R                  4S jr	Sr
U =r$ )ConvNextEmbeddingsj   zThis class is comparable to (and inspired by) the SwinEmbeddings class
found in src/transformers/models/swin/modeling_swin.py.
c                   > [         TU ]  5         [        R                  " UR                  UR
                  S   UR                  UR                  S9U l        [        UR
                  S   SSS9U l	        UR                  U l        g )Nr   kernel_sizestriderf   rM   rS   rT   )
r/   r0   r   Conv2dnum_channelshidden_sizes
patch_sizepatch_embeddingsrI   	layernormr1   configr2   s     r'   r0   ConvNextEmbeddings.__init__o   sr     "		!4!4Q!7VEVEV_e_p_p!
 +6+>+>q+AtYij"//r)   pixel_valuesr   c                     UR                   S   nX R                  :w  a  [        S5      eU R                  U5      nU R	                  U5      nU$ )Nr   zeMake sure that the channel dimension of the pixel values match with the one set in the configuration.)r   rp   
ValueErrorrs   rt   )r1   rx   rp   
embeddingss       r'   r7   ConvNextEmbeddings.forwardw   sT    #))!,,,,w  **<8
^^J/
r)   )rt   rp   rs   r=   r>   r?   r@   rA   r0   r    FloatTensorrC   r7   rE   rF   rG   s   @r'   rh   rh   j   s/    0E$5$5 %,,  r)   rh   c                   j   ^  \ rS rSrSrSU 4S jjrS\R                  S\R                  4S jr	Sr
U =r$ )	ConvNextLayer   a  This corresponds to the `Block` class in the original implementation.

There are two equivalent implementations: [DwConv, LayerNorm (channels_first), Conv, GELU,1x1 Conv]; all in (N, C,
H, W) (2) [DwConv, Permute to (N, H, W, C), LayerNorm (channels_last), Linear, GELU, Linear]; Permute back

The authors used (2) as they find it slightly faster in PyTorch.

Args:
    config ([`ConvNextConfig`]): Model configuration class.
    dim (`int`): Number of input channels.
    drop_path (`float`): Stochastic depth rate. Default: 0.0.
c                    > [         TU ]  5         [        R                  " X"SSUS9U l        [        USS9U l        [        R                  " USU-  5      U l        [        UR                     U l        [        R                  " SU-  U5      U l        UR                  S:  a6  [        R                  " UR                  [        R                   " U5      -  SS	9OS U l        US
:  a  [%        U5      U l        g [        R&                  " 5       U l        g )N   r
   )rl   paddinggroupsrf   rS      r   T)requires_gradr   )r/   r0   r   ro   dwconvrI   rt   Linearpwconv1r   
hidden_actactpwconv2layer_scale_init_valuerN   r    rO   layer_scale_parameterr+   Identityr(   )r1   rv   dimr(   r2   s       r'   r0   ConvNextLayer.__init__   s    iia3O*3D9yya#g.&++,yyS#. ,,q0 LL66S9JJZ^_ 	"
 9BC))4R[[]r)   r4   r   c                 b   UnU R                  U5      nUR                  SSSS5      nU R                  U5      nU R                  U5      nU R	                  U5      nU R                  U5      nU R                  b  U R                  U-  nUR                  SSSS5      nX R                  U5      -   nU$ )Nr   r[   r
   r   )r   permutert   r   r   r   r   r(   )r1   r4   r   rX   s       r'   r7   ConvNextLayer.forward   s    KK&IIaAq!NN1LLOHHQKLLO%%1**Q.AIIaAq!NN1%%r)   )r   r(   r   r   rt   r   r   )r   r}   rG   s   @r'   r   r      s0    [U%6%6 5<<  r)   r   c                   j   ^  \ rS rSrSrSU 4S jjrS\R                  S\R                  4S jr	Sr
U =r$ )	ConvNextStage   a}  ConvNeXT stage, consisting of an optional downsampling layer + multiple residual blocks.

Args:
    config ([`ConvNextConfig`]): Model configuration class.
    in_channels (`int`): Number of input channels.
    out_channels (`int`): Number of output channels.
    depth (`int`): Number of residual blocks.
    drop_path_rates(`List[float]`): Stochastic depth rates for each layer.
c                 ~  > [         T	U ]  5         X#:w  d  US:  a9  [        R                  " [	        USSS9[        R
                  " X#XES95      U l        O[        R                  " 5       U l        U=(       d    S/U-  n[        R                  " [        U5       Vs/ s H  n[        XXx   S9PM     sn6 U l
        g s  snf )Nr   rf   rM   rn   rk   r   )r   r(   )r/   r0   r   
SequentialrI   ro   downsampling_layerr   ranger   layers)
r1   rv   in_channelsout_channelsrl   rm   depthdrop_path_ratesjr2   s
            r'   r0   ConvNextStage.__init__   s    &&1*&(mm!+4EUV		+\'D#
 ')kkmD#):cUU]mm]bch]ij]iXYmF@RS]ij
js   B:r4   r   c                 J    U R                  U5      nU R                  U5      nU$ r.   r   r   r6   s     r'   r7   ConvNextStage.forward   s&    //>M2r)   r   )r[   r[   r[   Nr}   rG   s   @r'   r   r      s/    
U%6%6 5<<  r)   r   c                   t   ^  \ rS rSrU 4S jr  S	S\R                  S\\   S\\   S\	\
\4   4S jjrSrU =r$ )
ConvNextEncoder   c           
      .  > [         TU ]  5         [        R                  " 5       U l        [
        R                  " SUR                  [        UR                  5      SS9R                  UR                  5       Vs/ s H  nUR                  5       PM     nnUR                  S   n[        UR                  5       HT  nUR                  U   n[        UUUUS:  a  SOSUR                  U   X5   S9nU R                  R!                  U5        UnMV     g s  snf )Nr   cpu)r   r[   r   )r   r   rm   r   r   )r/   r0   r   
ModuleListstagesr    linspacedrop_path_ratesumdepthssplittolistrq   r   
num_stagesr   append)	r1   rv   rX   r   prev_chsiout_chsstager2   s	           r'   r0   ConvNextEncoder.__init__   s    mmo ^^Av'<'<c&-->PY^_eeflfsfst
t HHJt 	 
 &&q)v(()A))!,G!$$EqqmmA& / 2E KKu%H *
s   9Dr4   output_hidden_statesreturn_dictr   c                     U(       a  SOS n[        U R                  5       H  u  pVU(       a  XA4-   nU" U5      nM     U(       a  XA4-   nU(       d  [        S X4 5       5      $ [        UUS9$ )N c              3   .   #    U  H  oc  M  Uv   M     g 7fr.   r   ).0vs     r'   	<genexpr>*ConvNextEncoder.forward.<locals>.<genexpr>   s     X$Fq$Fs   	)last_hidden_stater4   )	enumerater   tupler   )r1   r4   r   r   all_hidden_statesr   layer_modules          r'   r7   ConvNextEncoder.forward   sw     #7BD(5OA#$58H$H!(7M	  6   14D DX]$FXXX-++
 	
r)   )r   )FT)r=   r>   r?   r@   r0   r    r~   r   boolr   r   r   r7   rE   rF   rG   s   @r'   r   r      sY    0 05&*	
((
 'tn
 d^	

 
u44	5
 
r)   r   c                   ,    \ rS rSr\rSrSrS/rS r	Sr
g)ConvNextPreTrainedModel   convnextrx   r   c                    [        U[        R                  [        R                  45      (       ak  UR                  R
                  R                  SU R                  R                  S9  UR                  b%  UR                  R
                  R                  5         gg[        U[        R                  [        45      (       aJ  UR                  R
                  R                  5         UR                  R
                  R                  S5        g[        U[        5      (       aH  UR                  b:  UR                  R
                  R                  U R                  R                   5        ggg)zInitialize the weightsr   )r^   stdNg      ?)
isinstancer   r   ro   rP   datanormal_rv   initializer_rangerR   zero_	LayerNormrI   fill_r   r   r   )r1   modules     r'   _init_weights%ConvNextPreTrainedModel._init_weights  s    fryy"))455 MM&&CT[[5R5R&S{{&  &&( '/@ ABBKK""$MM$$S)..++7,,11778Z8Z[ 8 /r)   r   N)r=   r>   r?   r@   r   config_classbase_model_prefixmain_input_name_no_split_modulesr   rE   r   r)   r'   r   r      s!    !L"$O()\r)   r   c                      ^  \ rS rSrU 4S jr\   S	S\\R                     S\\	   S\\	   S\
\\4   4S jj5       rSrU =r$ )
ConvNextModeli  c                    > [         TU ]  U5        Xl        [        U5      U l        [        U5      U l        [        R                  " UR                  S   UR                  S9U l        U R                  5         g )Nr   )r/   r0   rv   rh   r{   r   encoderr   r   rq   layer_norm_epsrt   	post_initru   s     r'   r0   ConvNextModel.__init__  s^     ,V4&v. f&9&9"&=6CXCXY 	r)   rx   r   r   r   c                 f   Ub  UOU R                   R                  nUb  UOU R                   R                  nUc  [        S5      eU R	                  U5      nU R                  UUUS9nUS   nU R                  UR                  SS/5      5      nU(       d	  Xg4USS  -   $ [        UUUR                  S9$ )Nz You have to specify pixel_valuesr   r   r   r   r   )r   pooler_outputr4   )
rv   r   use_return_dictrz   r{   r   rt   r^   r   r4   )r1   rx   r   r   embedding_outputencoder_outputsr   pooled_outputs           r'   r7   ConvNextModel.forward$  s     %9$D $++JjJj 	 &1%<k$++B]B]?@@??<8,,!5# ' 
 ,A. '8'='=r2h'GH%58KKK7/')77
 	
r)   )rv   r{   r   rt   )NNN)r=   r>   r?   r@   r0   r   r   r    r~   r   r   r   r   r7   rE   rF   rG   s   @r'   r   r     sk      59/3&*	"
u001"
 'tn"
 d^	"

 
u>>	?"
 "
r)   r   z
    ConvNext Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for
    ImageNet.
    )custom_introc                      ^  \ rS rSrU 4S jr\    S
S\\R                     S\\R                     S\\
   S\\
   S\\\4   4
S jj5       rS	rU =r$ )ConvNextForImageClassificationiJ  c                 6  > [         TU ]  U5        UR                  U l        [        U5      U l        UR                  S:  a.  [
        R                  " UR                  S   UR                  5      O[
        R                  " 5       U l	        U R                  5         g )Nr   r   )r/   r0   
num_labelsr   r   r   r   rq   r   
classifierr   ru   s     r'   r0   'ConvNextForImageClassification.__init__Q  sy      ++%f- FLEVEVYZEZBIIf))"-v/@/@A`b`k`k`m 	
 	r)   rx   labelsr   r   r   c                 2   Ub  UOU R                   R                  nU R                  XUS9nU(       a  UR                  OUS   nU R	                  U5      nSnUGb  U R                   R
                  c  U R                  S:X  a  SU R                   l        OoU R                  S:  aN  UR                  [        R                  :X  d  UR                  [        R                  :X  a  SU R                   l        OSU R                   l        U R                   R
                  S:X  aI  [        5       n	U R                  S:X  a&  U	" UR                  5       UR                  5       5      nOU	" Xr5      nOU R                   R
                  S:X  a=  [        5       n	U	" UR                  SU R                  5      UR                  S5      5      nO,U R                   R
                  S:X  a  [        5       n	U	" Xr5      nU(       d  U4USS -   n
Ub  U4U
-   $ U
$ [!        UUUR"                  S	9$ )
ab  
labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*):
    Labels for computing the image classification/regression loss. Indices should be in `[0, ...,
    config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If
    `config.num_labels > 1` a classification loss is computed (Cross-Entropy).
Nr   r   
regressionsingle_label_classificationmulti_label_classificationr   r[   )losslogitsr4   )rv   r   r   r   r   problem_typer   r   r    longintr	   squeezer   viewr   r   r4   )r1   rx   r   r   r   outputsr   r   r   loss_fctr&   s              r'   r7   &ConvNextForImageClassification.forward_  s    &1%<k$++B]B]--ep-q1<--'!*/{{''/??a'/;DKK,__q(fllejj.HFLL\a\e\eLe/LDKK,/KDKK,{{''<7"9??a'#FNN$4fnn6FGD#F3D))-JJ+-B @&++b/R))-II,./Y,F)-)9TGf$EvE3!//
 	
r)   )r   r   r   )NNNN)r=   r>   r?   r@   r0   r   r   r    r~   
LongTensorr   r   r   r   r7   rE   rF   rG   s   @r'   r   r   J  s      59-1/3&*3
u0013
 ))*3
 'tn	3

 d^3
 
u::	;3
 3
r)   r   zQ
    ConvNeXt backbone, to be used with frameworks like DETR and MaskFormer.
    c            
       t   ^  \ rS rSrU 4S jr\  S	S\R                  S\\	   S\\	   S\
4S jj5       rSrU =r$ )
ConvNextBackbonei  c                   > [         TU ]  U5        [         TU ]	  U5        [        U5      U l        [        U5      U l        UR                  S   /UR                  -   U l        0 n[        U R                  U R                  5       H  u  p4[        USS9X#'   M     [        R                  " U5      U l        U R!                  5         g )Nr   rM   )rT   )r/   r0   _init_backbonerh   r{   r   r   rq   num_featureszip_out_featureschannelsrI   r   
ModuleDicthidden_states_normsr   )r1   rv   r  r   rp   r2   s        r'   r0   ConvNextBackbone.__init__  s     v&,V4&v.#0034v7J7JJ !#&t'9'94==#IE):<Ue)f& $J#%==1D#E  	r)   rx   r   r   r   c                    Ub  UOU R                   R                  nUb  UOU R                   R                  nU R                  U5      nU R	                  USUS9nU(       a  UR
                  OUS   nSn[        U R                  U5       H0  u  pXR                  ;   d  M  U R                  U   " U	5      n	Xy4-  nM2     U(       d  U4n
U(       a  X4-  n
U
$ [        UU(       a  USS9$ SSS9$ )a   
Examples:

```python
>>> from transformers import AutoImageProcessor, AutoBackbone
>>> import torch
>>> from PIL import Image
>>> import requests

>>> url = "http://images.cocodataset.org/val2017/000000039769.jpg"
>>> image = Image.open(requests.get(url, stream=True).raw)

>>> processor = AutoImageProcessor.from_pretrained("facebook/convnext-tiny-224")
>>> model = AutoBackbone.from_pretrained("facebook/convnext-tiny-224")

>>> inputs = processor(image, return_tensors="pt")
>>> outputs = model(**inputs)
```NTr   r   r   )feature_mapsr4   
attentions)rv   r   r   r{   r   r4   r  stage_namesout_featuresr  r   )r1   rx   r   r   r   r  r4   r  r   hidden_stater&   s              r'   r7   ConvNextBackbone.forward  s   2 &1%<k$++B]B]$8$D $++JjJj 	  ??<8,,!%#  
 2=--'!*#&t'7'7#GE)))#77>|L/ $H
 "_F#**M%+?-
 	
EI
 	
r)   )r{   r   r  r  )NN)r=   r>   r?   r@   r0   r   r    rC   r   r   r   r7   rE   rF   rG   s   @r'   r	  r	    sV    "  04&*	7
ll7
 'tn7
 d^	7

 
7
 7
r)   r	  )r   r   r   r	  )r   F)/rA   typingr   r   r   r    torch.utils.checkpointr   torch.nnr   r   r	   activationsr   modeling_outputsr   r   r   r   modeling_utilsr   utilsr   r   utils.backbone_utilsr   configuration_convnextr   
get_loggerr=   loggerrC   rB   r   r(   Moduler+   rI   rh   r   r   r   r   r   r   r	  __all__r   r)   r'   <module>r'     st    ) )    A A !  . , 1 2 
		H	%U\\ e T V[VbVb *-ryy -		 < 0)BII )XBII @-
bii -
` \o \ \, 1
+ 1
 1
h C
%< C
C
L 
J
. J

J
Z mr)   