receive incoming errors / 'rcv_fifo_overrun'

Showing results for 
Search instead for 
Did you mean: 

receive incoming errors / 'rcv_fifo_overrun'

L0 Member

Hello everyone. 


Got a question to the community on an interesting situation: 


- PAN-PA-3050 / PAN OS 7.1.10

- Drop Counter increases on two aggregated interfaces (ae3 - interfaces 1/3 & 1/4)

- connected via Cisco vPC tech with a Nexus FEX switch

- new patch cable

- only req. VLANs are on the trunk / CDP is deactivated

- no obvious layer 2 errors visible during packet capture


Counter details:


REDACTED (active)> show  interface ethernet1/3


Hardware interface counters read from CPU:


bytes received                           2198604878

bytes transmitted                        67193368

packets received                         16782405

packets transmitted                      541882

receive incoming errors                  1817713

receive discarded                        0

receive errors                           0

packets dropped                          0



REDACTED (active)> show interface ethernet1/4

Hardware interface counters read from CPU:


bytes received                           2198281424

bytes transmitted                        67192748

packets received                         16779807

packets transmitted                      541877

receive incoming errors                  1682781

receive discarded                        1

receive errors                           0

packets dropped                          0



REDACTED (active)> show system state filter sys.s1.p3.detail


sys.s1.p3.detail: { 'pkts1024tomax_octets': 0x73619d2bd, 'pkts128to255_octets': 0x7d82ed54, 'pkts256to511_octets': 0x2f804d10, 'pkts512to1023_octets': 0x4dd41a61, 'pkts64_octets': 0x21a793bf, 'pkts65to127_octe

ts': 0x2a3d76b7d, 'rcv_fifo_overrun': 0x1bbc71, }


REDACTED (active)> show system state filter sys.s1.p4.detail


sys.s1.p4.detail: { 'bad_crc': 0x1, 'pkts1024tomax_octets': 0x769a235ec, 'pkts128to255_octets': 0xe38e6000, 'pkts256to511_octets': 0x3106dc4a, 'pkts512to1023_octets': 0x4d595c47, 'pkts64_octets': 0x19c1a2b5, '

pkts65to127_octets': 0x284dd797b, 'rcv_fifo_overrun': 0x19ad5d, }



REDACTED (active)> show counter global filter delta yes severity drop


Global counters:

Elapsed time since last sampling: 12.214 seconds


name                                   value     rate severity  category  aspect    description


flow_rcv_dot1q_tag_err                    55        4 drop      flow      parse     Packets dropped: 802.1q tag not configured

flow_no_interface                         55        4 drop      flow      parse     Packets dropped: invalid interface

flow_ipv6_disabled                        14        1 drop      flow      parse     Packets dropped: IPv6 disabled on interface

flow_policy_deny                        1684      137 drop      flow      session   Session setup: denied by policy

flow_tcp_non_syn_drop                     20        1 drop      flow      session   Packets dropped: non-SYN TCP without session match

flow_fwd_l3_bcast_drop                     1        0 drop      flow      forward   Packets dropped: unhandled IP broadcast

flow_fwd_l3_mcast_drop                   206       16 drop      flow      forward   Packets dropped: no route for IP multicast

flow_fwd_notopology                        3        0 drop      flow      forward   Packets dropped: no forwarding configured on interface


Total counters shown: 8



We do not encounter any kind of issues within the network, but this is pretty confusing. Is this a bug or a normal behaviour? 

I might be mistaken, but does 'rcv_fifo_overrun' mean that the firewall cannot process the incoming traffic? This is almost impossible, since we do not experience connection issues. Is this maybe flood protection? Or am I missing something? 


Any input would be appreciated. 


L2 Linker

Same issue "receive fifo overrun"
even there's no clear document about this counter and how it increases ?

Hi vmilan,

Can I ask what is the model of the device you have problem with?



I have 2 different cases with different customer


One is in PA-850, and another is in PA-3060


In salesforce i found one of the case where enabling flow control both the side of the link resolved this issue, i tried in one of my case but doesn't work.


Overruns occure when the  buffer of the interface is full, but it is still trying to handle incoming traffic. The port has exceeded its buffer. This is usually due to bursty traffic more than anything else. 

Customer don't want this counters to increase....

Is there any way to reduce this counters ??



Depends on the platform and what port we're talking about. If you can upgrade the interface to a faster SFP module than yes, if not or if this is an ethernet interface than no. Essentially you need to figure out if bursty traffic is expected on the interface in question, if not then look into why it's present. If bursty traffic would be expected, than there isn't much you'll be able to do. 

Hi BPry,


Yes, i believe this is what i wanted to know and it will be helpful to convience customer.


And hopefully it will resolve this issue.


Thank you very much for your words @BPry @EliSz @Martin-Kowalski


If i need any further help, i will let you know.



Milan Variya

Thanks for so Valuable info.

Lot to learn from your posts.


L1 Bithead


We had similar case for 3 devices, problem with rcv_fifo_overrun counters increasing on external interface directly connected to the ISP router. 

Devices were on 8.1.5 and 8.0.12 PANOS, model PA820.

First we thought problem might be due to high packet burst where device can't handle it properly and simply drops the packets, also we suspected PAN-104116 leak or another thought was that the device could be simply to busy and drops the packets that can't handle.

Leak was fixed in 8.1.5 and we try the upgrade path, which didn’t help in the end.

Then when we noticed that 2 more devices (same model PA820) had the same problem, we have logged the case with PA TAC.

Upon troubleshooting engineer provided some more insight based on the past experience:

Data receive buffer overrun - is only indication of ingress traffic bursts on the marvell switch port(Firewall DP Ports) .rcv_fifo_overrun errors are suggesting that Marvell(PA's switch backplane) was experiencing buffer shortage/switch port on Marvell fabric is receiving more frames than can be managed and this can be related to bursty traffic pattern.


Errors can only be found on Marvell external ports (phy ports, RJ45) and not between internal

chipsets(From Marvell ---> Dataplane processor). Buffer size cannot be changed/adjusted.

> debug dataplane internal pdt marvell stats



What we are interested in is the UL1 and UL2 (Uplinks to Octeon Processor) and

they appear to be good, in all of them indicating marvell is not sending pause frames or

any recv errors caused by hardware.


 Also the traffic rate in my case and for that particular device was above 20K pps Rcv_fifo_overrun is expected/normal during bursty traffic pattern and it's not an issue on the firewall.


Problem is only when the traffic rate is low and if this counter increases, even with reasonable amount of traffic, then it needs investigation.


Also this doesn't have relevance to packet-descriptor depletion or leak as I have thought originally with reference to PAN-104116.

Because PD is a DP component and "rcv_fifo_overrun"  is drop stage in marvell even before it makes it to DP.


In my case from TAC perspective there is not much we can do.


Kind Regards,



Like what you see?

Show your appreciation!

Click Like if a post is helpful to you or if you just want to show your support.

Click Accept as Solution to acknowledge that the answer to your question has been provided.

The button appears next to the replies on topics you’ve started. The member who gave the solution and all future visitors to this topic will appreciate it!

These simple actions take just seconds of your time, but go a long way in showing appreciation for community members and the LIVEcommunity as a whole!

The LIVEcommunity thanks you for your participation!