spectrum_router.c 195 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223222422252226222722282229223022312232223322342235223622372238223922402241224222432244224522462247224822492250225122522253225422552256225722582259226022612262226322642265226622672268226922702271227222732274227522762277227822792280228122822283228422852286228722882289229022912292229322942295229622972298229923002301230223032304230523062307230823092310231123122313231423152316231723182319232023212322232323242325232623272328232923302331233223332334233523362337233823392340234123422343234423452346234723482349235023512352235323542355235623572358235923602361236223632364236523662367236823692370237123722373237423752376237723782379238023812382238323842385238623872388238923902391239223932394239523962397239823992400240124022403240424052406240724082409241024112412241324142415241624172418241924202421242224232424242524262427242824292430243124322433243424352436243724382439244024412442244324442445244624472448244924502451245224532454245524562457245824592460246124622463246424652466246724682469247024712472247324742475247624772478247924802481248224832484248524862487248824892490249124922493249424952496249724982499250025012502250325042505250625072508250925102511251225132514251525162517251825192520252125222523252425252526252725282529253025312532253325342535253625372538253925402541254225432544254525462547254825492550255125522553255425552556255725582559256025612562256325642565256625672568256925702571257225732574257525762577257825792580258125822583258425852586258725882589259025912592259325942595259625972598259926002601260226032604260526062607260826092610261126122613261426152616261726182619262026212622262326242625262626272628262926302631263226332634263526362637263826392640264126422643264426452646264726482649265026512652265326542655265626572658265926602661266226632664266526662667266826692670267126722673267426752676267726782679268026812682268326842685268626872688268926902691269226932694269526962697269826992700270127022703270427052706270727082709271027112712271327142715271627172718271927202721272227232724272527262727272827292730273127322733273427352736273727382739274027412742274327442745274627472748274927502751275227532754275527562757275827592760276127622763276427652766276727682769277027712772277327742775277627772778277927802781278227832784278527862787278827892790279127922793279427952796279727982799280028012802280328042805280628072808280928102811281228132814281528162817281828192820282128222823282428252826282728282829283028312832283328342835283628372838283928402841284228432844284528462847284828492850285128522853285428552856285728582859286028612862286328642865286628672868286928702871287228732874287528762877287828792880288128822883288428852886288728882889289028912892289328942895289628972898289929002901290229032904290529062907290829092910291129122913291429152916291729182919292029212922292329242925292629272928292929302931293229332934293529362937293829392940294129422943294429452946294729482949295029512952295329542955295629572958295929602961296229632964296529662967296829692970297129722973297429752976297729782979298029812982298329842985298629872988298929902991299229932994299529962997299829993000300130023003300430053006300730083009301030113012301330143015301630173018301930203021302230233024302530263027302830293030303130323033303430353036303730383039304030413042304330443045304630473048304930503051305230533054305530563057305830593060306130623063306430653066306730683069307030713072307330743075307630773078307930803081308230833084308530863087308830893090309130923093309430953096309730983099310031013102310331043105310631073108310931103111311231133114311531163117311831193120312131223123312431253126312731283129313031313132313331343135313631373138313931403141314231433144314531463147314831493150315131523153315431553156315731583159316031613162316331643165316631673168316931703171317231733174317531763177317831793180318131823183318431853186318731883189319031913192319331943195319631973198319932003201320232033204320532063207320832093210321132123213321432153216321732183219322032213222322332243225322632273228322932303231323232333234323532363237323832393240324132423243324432453246324732483249325032513252325332543255325632573258325932603261326232633264326532663267326832693270327132723273327432753276327732783279328032813282328332843285328632873288328932903291329232933294329532963297329832993300330133023303330433053306330733083309331033113312331333143315331633173318331933203321332233233324332533263327332833293330333133323333333433353336333733383339334033413342334333443345334633473348334933503351335233533354335533563357335833593360336133623363336433653366336733683369337033713372337333743375337633773378337933803381338233833384338533863387338833893390339133923393339433953396339733983399340034013402340334043405340634073408340934103411341234133414341534163417341834193420342134223423342434253426342734283429343034313432343334343435343634373438343934403441344234433444344534463447344834493450345134523453345434553456345734583459346034613462346334643465346634673468346934703471347234733474347534763477347834793480348134823483348434853486348734883489349034913492349334943495349634973498349935003501350235033504350535063507350835093510351135123513351435153516351735183519352035213522352335243525352635273528352935303531353235333534353535363537353835393540354135423543354435453546354735483549355035513552355335543555355635573558355935603561356235633564356535663567356835693570357135723573357435753576357735783579358035813582358335843585358635873588358935903591359235933594359535963597359835993600360136023603360436053606360736083609361036113612361336143615361636173618361936203621362236233624362536263627362836293630363136323633363436353636363736383639364036413642364336443645364636473648364936503651365236533654365536563657365836593660366136623663366436653666366736683669367036713672367336743675367636773678367936803681368236833684368536863687368836893690369136923693369436953696369736983699370037013702370337043705370637073708370937103711371237133714371537163717371837193720372137223723372437253726372737283729373037313732373337343735373637373738373937403741374237433744374537463747374837493750375137523753375437553756375737583759376037613762376337643765376637673768376937703771377237733774377537763777377837793780378137823783378437853786378737883789379037913792379337943795379637973798379938003801380238033804380538063807380838093810381138123813381438153816381738183819382038213822382338243825382638273828382938303831383238333834383538363837383838393840384138423843384438453846384738483849385038513852385338543855385638573858385938603861386238633864386538663867386838693870387138723873387438753876387738783879388038813882388338843885388638873888388938903891389238933894389538963897389838993900390139023903390439053906390739083909391039113912391339143915391639173918391939203921392239233924392539263927392839293930393139323933393439353936393739383939394039413942394339443945394639473948394939503951395239533954395539563957395839593960396139623963396439653966396739683969397039713972397339743975397639773978397939803981398239833984398539863987398839893990399139923993399439953996399739983999400040014002400340044005400640074008400940104011401240134014401540164017401840194020402140224023402440254026402740284029403040314032403340344035403640374038403940404041404240434044404540464047404840494050405140524053405440554056405740584059406040614062406340644065406640674068406940704071407240734074407540764077407840794080408140824083408440854086408740884089409040914092409340944095409640974098409941004101410241034104410541064107410841094110411141124113411441154116411741184119412041214122412341244125412641274128412941304131413241334134413541364137413841394140414141424143414441454146414741484149415041514152415341544155415641574158415941604161416241634164416541664167416841694170417141724173417441754176417741784179418041814182418341844185418641874188418941904191419241934194419541964197419841994200420142024203420442054206420742084209421042114212421342144215421642174218421942204221422242234224422542264227422842294230423142324233423442354236423742384239424042414242424342444245424642474248424942504251425242534254425542564257425842594260426142624263426442654266426742684269427042714272427342744275427642774278427942804281428242834284428542864287428842894290429142924293429442954296429742984299430043014302430343044305430643074308430943104311431243134314431543164317431843194320432143224323432443254326432743284329433043314332433343344335433643374338433943404341434243434344434543464347434843494350435143524353435443554356435743584359436043614362436343644365436643674368436943704371437243734374437543764377437843794380438143824383438443854386438743884389439043914392439343944395439643974398439944004401440244034404440544064407440844094410441144124413441444154416441744184419442044214422442344244425442644274428442944304431443244334434443544364437443844394440444144424443444444454446444744484449445044514452445344544455445644574458445944604461446244634464446544664467446844694470447144724473447444754476447744784479448044814482448344844485448644874488448944904491449244934494449544964497449844994500450145024503450445054506450745084509451045114512451345144515451645174518451945204521452245234524452545264527452845294530453145324533453445354536453745384539454045414542454345444545454645474548454945504551455245534554455545564557455845594560456145624563456445654566456745684569457045714572457345744575457645774578457945804581458245834584458545864587458845894590459145924593459445954596459745984599460046014602460346044605460646074608460946104611461246134614461546164617461846194620462146224623462446254626462746284629463046314632463346344635463646374638463946404641464246434644464546464647464846494650465146524653465446554656465746584659466046614662466346644665466646674668466946704671467246734674467546764677467846794680468146824683468446854686468746884689469046914692469346944695469646974698469947004701470247034704470547064707470847094710471147124713471447154716471747184719472047214722472347244725472647274728472947304731473247334734473547364737473847394740474147424743474447454746474747484749475047514752475347544755475647574758475947604761476247634764476547664767476847694770477147724773477447754776477747784779478047814782478347844785478647874788478947904791479247934794479547964797479847994800480148024803480448054806480748084809481048114812481348144815481648174818481948204821482248234824482548264827482848294830483148324833483448354836483748384839484048414842484348444845484648474848484948504851485248534854485548564857485848594860486148624863486448654866486748684869487048714872487348744875487648774878487948804881488248834884488548864887488848894890489148924893489448954896489748984899490049014902490349044905490649074908490949104911491249134914491549164917491849194920492149224923492449254926492749284929493049314932493349344935493649374938493949404941494249434944494549464947494849494950495149524953495449554956495749584959496049614962496349644965496649674968496949704971497249734974497549764977497849794980498149824983498449854986498749884989499049914992499349944995499649974998499950005001500250035004500550065007500850095010501150125013501450155016501750185019502050215022502350245025502650275028502950305031503250335034503550365037503850395040504150425043504450455046504750485049505050515052505350545055505650575058505950605061506250635064506550665067506850695070507150725073507450755076507750785079508050815082508350845085508650875088508950905091509250935094509550965097509850995100510151025103510451055106510751085109511051115112511351145115511651175118511951205121512251235124512551265127512851295130513151325133513451355136513751385139514051415142514351445145514651475148514951505151515251535154515551565157515851595160516151625163516451655166516751685169517051715172517351745175517651775178517951805181518251835184518551865187518851895190519151925193519451955196519751985199520052015202520352045205520652075208520952105211521252135214521552165217521852195220522152225223522452255226522752285229523052315232523352345235523652375238523952405241524252435244524552465247524852495250525152525253525452555256525752585259526052615262526352645265526652675268526952705271527252735274527552765277527852795280528152825283528452855286528752885289529052915292529352945295529652975298529953005301530253035304530553065307530853095310531153125313531453155316531753185319532053215322532353245325532653275328532953305331533253335334533553365337533853395340534153425343534453455346534753485349535053515352535353545355535653575358535953605361536253635364536553665367536853695370537153725373537453755376537753785379538053815382538353845385538653875388538953905391539253935394539553965397539853995400540154025403540454055406540754085409541054115412541354145415541654175418541954205421542254235424542554265427542854295430543154325433543454355436543754385439544054415442544354445445544654475448544954505451545254535454545554565457545854595460546154625463546454655466546754685469547054715472547354745475547654775478547954805481548254835484548554865487548854895490549154925493549454955496549754985499550055015502550355045505550655075508550955105511551255135514551555165517551855195520552155225523552455255526552755285529553055315532553355345535553655375538553955405541554255435544554555465547554855495550555155525553555455555556555755585559556055615562556355645565556655675568556955705571557255735574557555765577557855795580558155825583558455855586558755885589559055915592559355945595559655975598559956005601560256035604560556065607560856095610561156125613561456155616561756185619562056215622562356245625562656275628562956305631563256335634563556365637563856395640564156425643564456455646564756485649565056515652565356545655565656575658565956605661566256635664566556665667566856695670567156725673567456755676567756785679568056815682568356845685568656875688568956905691569256935694569556965697569856995700570157025703570457055706570757085709571057115712571357145715571657175718571957205721572257235724572557265727572857295730573157325733573457355736573757385739574057415742574357445745574657475748574957505751575257535754575557565757575857595760576157625763576457655766576757685769577057715772577357745775577657775778577957805781578257835784578557865787578857895790579157925793579457955796579757985799580058015802580358045805580658075808580958105811581258135814581558165817581858195820582158225823582458255826582758285829583058315832583358345835583658375838583958405841584258435844584558465847584858495850585158525853585458555856585758585859586058615862586358645865586658675868586958705871587258735874587558765877587858795880588158825883588458855886588758885889589058915892589358945895589658975898589959005901590259035904590559065907590859095910591159125913591459155916591759185919592059215922592359245925592659275928592959305931593259335934593559365937593859395940594159425943594459455946594759485949595059515952595359545955595659575958595959605961596259635964596559665967596859695970597159725973597459755976597759785979598059815982598359845985598659875988598959905991599259935994599559965997599859996000600160026003600460056006600760086009601060116012601360146015601660176018601960206021602260236024602560266027602860296030603160326033603460356036603760386039604060416042604360446045604660476048604960506051605260536054605560566057605860596060606160626063606460656066606760686069607060716072607360746075607660776078607960806081608260836084608560866087608860896090609160926093609460956096609760986099610061016102610361046105610661076108610961106111611261136114611561166117611861196120612161226123612461256126612761286129613061316132613361346135613661376138613961406141614261436144614561466147614861496150615161526153615461556156615761586159616061616162616361646165616661676168616961706171617261736174617561766177617861796180618161826183618461856186618761886189619061916192619361946195619661976198619962006201620262036204620562066207620862096210621162126213621462156216621762186219622062216222622362246225622662276228622962306231623262336234623562366237623862396240624162426243624462456246624762486249625062516252625362546255625662576258625962606261626262636264626562666267626862696270627162726273627462756276627762786279628062816282628362846285628662876288628962906291629262936294629562966297629862996300630163026303630463056306630763086309631063116312631363146315631663176318631963206321632263236324632563266327632863296330633163326333633463356336633763386339634063416342634363446345634663476348634963506351635263536354635563566357635863596360636163626363636463656366636763686369637063716372637363746375637663776378637963806381638263836384638563866387638863896390639163926393639463956396639763986399640064016402640364046405640664076408640964106411641264136414641564166417641864196420642164226423642464256426642764286429643064316432643364346435643664376438643964406441644264436444644564466447644864496450645164526453645464556456645764586459646064616462646364646465646664676468646964706471647264736474647564766477647864796480648164826483648464856486648764886489649064916492649364946495649664976498649965006501650265036504650565066507650865096510651165126513651465156516651765186519652065216522652365246525652665276528652965306531653265336534653565366537653865396540654165426543654465456546654765486549655065516552655365546555655665576558655965606561656265636564656565666567656865696570657165726573657465756576657765786579658065816582658365846585658665876588658965906591659265936594659565966597659865996600660166026603660466056606660766086609661066116612661366146615661666176618661966206621662266236624662566266627662866296630663166326633663466356636663766386639664066416642664366446645664666476648664966506651665266536654665566566657665866596660666166626663666466656666666766686669667066716672667366746675667666776678667966806681668266836684668566866687668866896690669166926693669466956696669766986699670067016702670367046705670667076708670967106711671267136714671567166717671867196720672167226723672467256726672767286729673067316732673367346735673667376738673967406741674267436744674567466747674867496750675167526753675467556756675767586759676067616762676367646765676667676768676967706771677267736774677567766777677867796780678167826783678467856786678767886789679067916792679367946795679667976798679968006801680268036804680568066807680868096810681168126813681468156816681768186819682068216822682368246825682668276828682968306831683268336834683568366837683868396840684168426843684468456846684768486849685068516852685368546855685668576858685968606861686268636864686568666867686868696870687168726873687468756876687768786879688068816882688368846885688668876888688968906891689268936894689568966897689868996900690169026903690469056906690769086909691069116912691369146915691669176918691969206921692269236924692569266927692869296930693169326933693469356936693769386939694069416942694369446945694669476948694969506951695269536954695569566957695869596960696169626963696469656966696769686969697069716972697369746975697669776978697969806981698269836984698569866987698869896990699169926993699469956996699769986999700070017002700370047005700670077008700970107011701270137014701570167017701870197020702170227023702470257026702770287029703070317032703370347035703670377038703970407041704270437044704570467047704870497050705170527053705470557056705770587059706070617062706370647065706670677068706970707071707270737074707570767077707870797080708170827083708470857086708770887089709070917092709370947095709670977098709971007101710271037104710571067107710871097110711171127113711471157116711771187119712071217122712371247125712671277128712971307131713271337134713571367137713871397140714171427143714471457146714771487149715071517152715371547155715671577158715971607161716271637164716571667167716871697170717171727173717471757176717771787179718071817182718371847185718671877188718971907191719271937194719571967197719871997200720172027203720472057206720772087209721072117212721372147215721672177218721972207221722272237224722572267227722872297230723172327233723472357236723772387239724072417242724372447245724672477248724972507251725272537254725572567257725872597260726172627263726472657266726772687269727072717272727372747275727672777278727972807281728272837284728572867287728872897290729172927293729472957296729772987299730073017302730373047305730673077308730973107311731273137314
  1. /*
  2. * drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
  3. * Copyright (c) 2016-2018 Mellanox Technologies. All rights reserved.
  4. * Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
  5. * Copyright (c) 2016 Ido Schimmel <idosch@mellanox.com>
  6. * Copyright (c) 2016 Yotam Gigi <yotamg@mellanox.com>
  7. * Copyright (c) 2017-2018 Petr Machata <petrm@mellanox.com>
  8. *
  9. * Redistribution and use in source and binary forms, with or without
  10. * modification, are permitted provided that the following conditions are met:
  11. *
  12. * 1. Redistributions of source code must retain the above copyright
  13. * notice, this list of conditions and the following disclaimer.
  14. * 2. Redistributions in binary form must reproduce the above copyright
  15. * notice, this list of conditions and the following disclaimer in the
  16. * documentation and/or other materials provided with the distribution.
  17. * 3. Neither the names of the copyright holders nor the names of its
  18. * contributors may be used to endorse or promote products derived from
  19. * this software without specific prior written permission.
  20. *
  21. * Alternatively, this software may be distributed under the terms of the
  22. * GNU General Public License ("GPL") version 2 as published by the Free
  23. * Software Foundation.
  24. *
  25. * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
  26. * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
  27. * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
  28. * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
  29. * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
  30. * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
  31. * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
  32. * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
  33. * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
  34. * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
  35. * POSSIBILITY OF SUCH DAMAGE.
  36. */
  37. #include <linux/kernel.h>
  38. #include <linux/types.h>
  39. #include <linux/rhashtable.h>
  40. #include <linux/bitops.h>
  41. #include <linux/in6.h>
  42. #include <linux/notifier.h>
  43. #include <linux/inetdevice.h>
  44. #include <linux/netdevice.h>
  45. #include <linux/if_bridge.h>
  46. #include <linux/socket.h>
  47. #include <linux/route.h>
  48. #include <linux/gcd.h>
  49. #include <linux/random.h>
  50. #include <net/netevent.h>
  51. #include <net/neighbour.h>
  52. #include <net/arp.h>
  53. #include <net/ip_fib.h>
  54. #include <net/ip6_fib.h>
  55. #include <net/fib_rules.h>
  56. #include <net/ip_tunnels.h>
  57. #include <net/l3mdev.h>
  58. #include <net/addrconf.h>
  59. #include <net/ndisc.h>
  60. #include <net/ipv6.h>
  61. #include <net/fib_notifier.h>
  62. #include "spectrum.h"
  63. #include "core.h"
  64. #include "reg.h"
  65. #include "spectrum_cnt.h"
  66. #include "spectrum_dpipe.h"
  67. #include "spectrum_ipip.h"
  68. #include "spectrum_mr.h"
  69. #include "spectrum_mr_tcam.h"
  70. #include "spectrum_router.h"
  71. #include "spectrum_span.h"
  72. struct mlxsw_sp_fib;
  73. struct mlxsw_sp_vr;
  74. struct mlxsw_sp_lpm_tree;
  75. struct mlxsw_sp_rif_ops;
  76. struct mlxsw_sp_router {
  77. struct mlxsw_sp *mlxsw_sp;
  78. struct mlxsw_sp_rif **rifs;
  79. struct mlxsw_sp_vr *vrs;
  80. struct rhashtable neigh_ht;
  81. struct rhashtable nexthop_group_ht;
  82. struct rhashtable nexthop_ht;
  83. struct list_head nexthop_list;
  84. struct {
  85. /* One tree for each protocol: IPv4 and IPv6 */
  86. struct mlxsw_sp_lpm_tree *proto_trees[2];
  87. struct mlxsw_sp_lpm_tree *trees;
  88. unsigned int tree_count;
  89. } lpm;
  90. struct {
  91. struct delayed_work dw;
  92. unsigned long interval; /* ms */
  93. } neighs_update;
  94. struct delayed_work nexthop_probe_dw;
  95. #define MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL 5000 /* ms */
  96. struct list_head nexthop_neighs_list;
  97. struct list_head ipip_list;
  98. bool aborted;
  99. struct notifier_block fib_nb;
  100. struct notifier_block netevent_nb;
  101. const struct mlxsw_sp_rif_ops **rif_ops_arr;
  102. const struct mlxsw_sp_ipip_ops **ipip_ops_arr;
  103. };
  104. struct mlxsw_sp_rif {
  105. struct list_head nexthop_list;
  106. struct list_head neigh_list;
  107. struct net_device *dev;
  108. struct mlxsw_sp_fid *fid;
  109. unsigned char addr[ETH_ALEN];
  110. int mtu;
  111. u16 rif_index;
  112. u16 vr_id;
  113. const struct mlxsw_sp_rif_ops *ops;
  114. struct mlxsw_sp *mlxsw_sp;
  115. unsigned int counter_ingress;
  116. bool counter_ingress_valid;
  117. unsigned int counter_egress;
  118. bool counter_egress_valid;
  119. };
  120. struct mlxsw_sp_rif_params {
  121. struct net_device *dev;
  122. union {
  123. u16 system_port;
  124. u16 lag_id;
  125. };
  126. u16 vid;
  127. bool lag;
  128. };
  129. struct mlxsw_sp_rif_subport {
  130. struct mlxsw_sp_rif common;
  131. union {
  132. u16 system_port;
  133. u16 lag_id;
  134. };
  135. u16 vid;
  136. bool lag;
  137. };
  138. struct mlxsw_sp_rif_ipip_lb {
  139. struct mlxsw_sp_rif common;
  140. struct mlxsw_sp_rif_ipip_lb_config lb_config;
  141. u16 ul_vr_id; /* Reserved for Spectrum-2. */
  142. };
  143. struct mlxsw_sp_rif_params_ipip_lb {
  144. struct mlxsw_sp_rif_params common;
  145. struct mlxsw_sp_rif_ipip_lb_config lb_config;
  146. };
  147. struct mlxsw_sp_rif_ops {
  148. enum mlxsw_sp_rif_type type;
  149. size_t rif_size;
  150. void (*setup)(struct mlxsw_sp_rif *rif,
  151. const struct mlxsw_sp_rif_params *params);
  152. int (*configure)(struct mlxsw_sp_rif *rif);
  153. void (*deconfigure)(struct mlxsw_sp_rif *rif);
  154. struct mlxsw_sp_fid * (*fid_get)(struct mlxsw_sp_rif *rif);
  155. };
  156. static void mlxsw_sp_lpm_tree_hold(struct mlxsw_sp_lpm_tree *lpm_tree);
  157. static void mlxsw_sp_lpm_tree_put(struct mlxsw_sp *mlxsw_sp,
  158. struct mlxsw_sp_lpm_tree *lpm_tree);
  159. static int mlxsw_sp_vr_lpm_tree_bind(struct mlxsw_sp *mlxsw_sp,
  160. const struct mlxsw_sp_fib *fib,
  161. u8 tree_id);
  162. static int mlxsw_sp_vr_lpm_tree_unbind(struct mlxsw_sp *mlxsw_sp,
  163. const struct mlxsw_sp_fib *fib);
  164. static unsigned int *
  165. mlxsw_sp_rif_p_counter_get(struct mlxsw_sp_rif *rif,
  166. enum mlxsw_sp_rif_counter_dir dir)
  167. {
  168. switch (dir) {
  169. case MLXSW_SP_RIF_COUNTER_EGRESS:
  170. return &rif->counter_egress;
  171. case MLXSW_SP_RIF_COUNTER_INGRESS:
  172. return &rif->counter_ingress;
  173. }
  174. return NULL;
  175. }
  176. static bool
  177. mlxsw_sp_rif_counter_valid_get(struct mlxsw_sp_rif *rif,
  178. enum mlxsw_sp_rif_counter_dir dir)
  179. {
  180. switch (dir) {
  181. case MLXSW_SP_RIF_COUNTER_EGRESS:
  182. return rif->counter_egress_valid;
  183. case MLXSW_SP_RIF_COUNTER_INGRESS:
  184. return rif->counter_ingress_valid;
  185. }
  186. return false;
  187. }
  188. static void
  189. mlxsw_sp_rif_counter_valid_set(struct mlxsw_sp_rif *rif,
  190. enum mlxsw_sp_rif_counter_dir dir,
  191. bool valid)
  192. {
  193. switch (dir) {
  194. case MLXSW_SP_RIF_COUNTER_EGRESS:
  195. rif->counter_egress_valid = valid;
  196. break;
  197. case MLXSW_SP_RIF_COUNTER_INGRESS:
  198. rif->counter_ingress_valid = valid;
  199. break;
  200. }
  201. }
  202. static int mlxsw_sp_rif_counter_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
  203. unsigned int counter_index, bool enable,
  204. enum mlxsw_sp_rif_counter_dir dir)
  205. {
  206. char ritr_pl[MLXSW_REG_RITR_LEN];
  207. bool is_egress = false;
  208. int err;
  209. if (dir == MLXSW_SP_RIF_COUNTER_EGRESS)
  210. is_egress = true;
  211. mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
  212. err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  213. if (err)
  214. return err;
  215. mlxsw_reg_ritr_counter_pack(ritr_pl, counter_index, enable,
  216. is_egress);
  217. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  218. }
  219. int mlxsw_sp_rif_counter_value_get(struct mlxsw_sp *mlxsw_sp,
  220. struct mlxsw_sp_rif *rif,
  221. enum mlxsw_sp_rif_counter_dir dir, u64 *cnt)
  222. {
  223. char ricnt_pl[MLXSW_REG_RICNT_LEN];
  224. unsigned int *p_counter_index;
  225. bool valid;
  226. int err;
  227. valid = mlxsw_sp_rif_counter_valid_get(rif, dir);
  228. if (!valid)
  229. return -EINVAL;
  230. p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
  231. if (!p_counter_index)
  232. return -EINVAL;
  233. mlxsw_reg_ricnt_pack(ricnt_pl, *p_counter_index,
  234. MLXSW_REG_RICNT_OPCODE_NOP);
  235. err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
  236. if (err)
  237. return err;
  238. *cnt = mlxsw_reg_ricnt_good_unicast_packets_get(ricnt_pl);
  239. return 0;
  240. }
  241. static int mlxsw_sp_rif_counter_clear(struct mlxsw_sp *mlxsw_sp,
  242. unsigned int counter_index)
  243. {
  244. char ricnt_pl[MLXSW_REG_RICNT_LEN];
  245. mlxsw_reg_ricnt_pack(ricnt_pl, counter_index,
  246. MLXSW_REG_RICNT_OPCODE_CLEAR);
  247. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ricnt), ricnt_pl);
  248. }
  249. int mlxsw_sp_rif_counter_alloc(struct mlxsw_sp *mlxsw_sp,
  250. struct mlxsw_sp_rif *rif,
  251. enum mlxsw_sp_rif_counter_dir dir)
  252. {
  253. unsigned int *p_counter_index;
  254. int err;
  255. p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
  256. if (!p_counter_index)
  257. return -EINVAL;
  258. err = mlxsw_sp_counter_alloc(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
  259. p_counter_index);
  260. if (err)
  261. return err;
  262. err = mlxsw_sp_rif_counter_clear(mlxsw_sp, *p_counter_index);
  263. if (err)
  264. goto err_counter_clear;
  265. err = mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
  266. *p_counter_index, true, dir);
  267. if (err)
  268. goto err_counter_edit;
  269. mlxsw_sp_rif_counter_valid_set(rif, dir, true);
  270. return 0;
  271. err_counter_edit:
  272. err_counter_clear:
  273. mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
  274. *p_counter_index);
  275. return err;
  276. }
  277. void mlxsw_sp_rif_counter_free(struct mlxsw_sp *mlxsw_sp,
  278. struct mlxsw_sp_rif *rif,
  279. enum mlxsw_sp_rif_counter_dir dir)
  280. {
  281. unsigned int *p_counter_index;
  282. if (!mlxsw_sp_rif_counter_valid_get(rif, dir))
  283. return;
  284. p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir);
  285. if (WARN_ON(!p_counter_index))
  286. return;
  287. mlxsw_sp_rif_counter_edit(mlxsw_sp, rif->rif_index,
  288. *p_counter_index, false, dir);
  289. mlxsw_sp_counter_free(mlxsw_sp, MLXSW_SP_COUNTER_SUB_POOL_RIF,
  290. *p_counter_index);
  291. mlxsw_sp_rif_counter_valid_set(rif, dir, false);
  292. }
  293. static void mlxsw_sp_rif_counters_alloc(struct mlxsw_sp_rif *rif)
  294. {
  295. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  296. struct devlink *devlink;
  297. devlink = priv_to_devlink(mlxsw_sp->core);
  298. if (!devlink_dpipe_table_counter_enabled(devlink,
  299. MLXSW_SP_DPIPE_TABLE_NAME_ERIF))
  300. return;
  301. mlxsw_sp_rif_counter_alloc(mlxsw_sp, rif, MLXSW_SP_RIF_COUNTER_EGRESS);
  302. }
  303. static void mlxsw_sp_rif_counters_free(struct mlxsw_sp_rif *rif)
  304. {
  305. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  306. mlxsw_sp_rif_counter_free(mlxsw_sp, rif, MLXSW_SP_RIF_COUNTER_EGRESS);
  307. }
  308. static struct mlxsw_sp_rif *
  309. mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
  310. const struct net_device *dev);
  311. #define MLXSW_SP_PREFIX_COUNT (sizeof(struct in6_addr) * BITS_PER_BYTE + 1)
  312. struct mlxsw_sp_prefix_usage {
  313. DECLARE_BITMAP(b, MLXSW_SP_PREFIX_COUNT);
  314. };
  315. #define mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) \
  316. for_each_set_bit(prefix, (prefix_usage)->b, MLXSW_SP_PREFIX_COUNT)
  317. static bool
  318. mlxsw_sp_prefix_usage_eq(struct mlxsw_sp_prefix_usage *prefix_usage1,
  319. struct mlxsw_sp_prefix_usage *prefix_usage2)
  320. {
  321. return !memcmp(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
  322. }
  323. static void
  324. mlxsw_sp_prefix_usage_cpy(struct mlxsw_sp_prefix_usage *prefix_usage1,
  325. struct mlxsw_sp_prefix_usage *prefix_usage2)
  326. {
  327. memcpy(prefix_usage1, prefix_usage2, sizeof(*prefix_usage1));
  328. }
  329. static void
  330. mlxsw_sp_prefix_usage_set(struct mlxsw_sp_prefix_usage *prefix_usage,
  331. unsigned char prefix_len)
  332. {
  333. set_bit(prefix_len, prefix_usage->b);
  334. }
  335. static void
  336. mlxsw_sp_prefix_usage_clear(struct mlxsw_sp_prefix_usage *prefix_usage,
  337. unsigned char prefix_len)
  338. {
  339. clear_bit(prefix_len, prefix_usage->b);
  340. }
  341. struct mlxsw_sp_fib_key {
  342. unsigned char addr[sizeof(struct in6_addr)];
  343. unsigned char prefix_len;
  344. };
  345. enum mlxsw_sp_fib_entry_type {
  346. MLXSW_SP_FIB_ENTRY_TYPE_REMOTE,
  347. MLXSW_SP_FIB_ENTRY_TYPE_LOCAL,
  348. MLXSW_SP_FIB_ENTRY_TYPE_TRAP,
  349. /* This is a special case of local delivery, where a packet should be
  350. * decapsulated on reception. Note that there is no corresponding ENCAP,
  351. * because that's a type of next hop, not of FIB entry. (There can be
  352. * several next hops in a REMOTE entry, and some of them may be
  353. * encapsulating entries.)
  354. */
  355. MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP,
  356. };
  357. struct mlxsw_sp_nexthop_group;
  358. struct mlxsw_sp_fib_node {
  359. struct list_head entry_list;
  360. struct list_head list;
  361. struct rhash_head ht_node;
  362. struct mlxsw_sp_fib *fib;
  363. struct mlxsw_sp_fib_key key;
  364. };
  365. struct mlxsw_sp_fib_entry_decap {
  366. struct mlxsw_sp_ipip_entry *ipip_entry;
  367. u32 tunnel_index;
  368. };
  369. struct mlxsw_sp_fib_entry {
  370. struct list_head list;
  371. struct mlxsw_sp_fib_node *fib_node;
  372. enum mlxsw_sp_fib_entry_type type;
  373. struct list_head nexthop_group_node;
  374. struct mlxsw_sp_nexthop_group *nh_group;
  375. struct mlxsw_sp_fib_entry_decap decap; /* Valid for decap entries. */
  376. };
  377. struct mlxsw_sp_fib4_entry {
  378. struct mlxsw_sp_fib_entry common;
  379. u32 tb_id;
  380. u32 prio;
  381. u8 tos;
  382. u8 type;
  383. };
  384. struct mlxsw_sp_fib6_entry {
  385. struct mlxsw_sp_fib_entry common;
  386. struct list_head rt6_list;
  387. unsigned int nrt6;
  388. };
  389. struct mlxsw_sp_rt6 {
  390. struct list_head list;
  391. struct fib6_info *rt;
  392. };
  393. struct mlxsw_sp_lpm_tree {
  394. u8 id; /* tree ID */
  395. unsigned int ref_count;
  396. enum mlxsw_sp_l3proto proto;
  397. unsigned long prefix_ref_count[MLXSW_SP_PREFIX_COUNT];
  398. struct mlxsw_sp_prefix_usage prefix_usage;
  399. };
  400. struct mlxsw_sp_fib {
  401. struct rhashtable ht;
  402. struct list_head node_list;
  403. struct mlxsw_sp_vr *vr;
  404. struct mlxsw_sp_lpm_tree *lpm_tree;
  405. enum mlxsw_sp_l3proto proto;
  406. };
  407. struct mlxsw_sp_vr {
  408. u16 id; /* virtual router ID */
  409. u32 tb_id; /* kernel fib table id */
  410. unsigned int rif_count;
  411. struct mlxsw_sp_fib *fib4;
  412. struct mlxsw_sp_fib *fib6;
  413. struct mlxsw_sp_mr_table *mr_table[MLXSW_SP_L3_PROTO_MAX];
  414. };
  415. static const struct rhashtable_params mlxsw_sp_fib_ht_params;
  416. static struct mlxsw_sp_fib *mlxsw_sp_fib_create(struct mlxsw_sp *mlxsw_sp,
  417. struct mlxsw_sp_vr *vr,
  418. enum mlxsw_sp_l3proto proto)
  419. {
  420. struct mlxsw_sp_lpm_tree *lpm_tree;
  421. struct mlxsw_sp_fib *fib;
  422. int err;
  423. lpm_tree = mlxsw_sp->router->lpm.proto_trees[proto];
  424. fib = kzalloc(sizeof(*fib), GFP_KERNEL);
  425. if (!fib)
  426. return ERR_PTR(-ENOMEM);
  427. err = rhashtable_init(&fib->ht, &mlxsw_sp_fib_ht_params);
  428. if (err)
  429. goto err_rhashtable_init;
  430. INIT_LIST_HEAD(&fib->node_list);
  431. fib->proto = proto;
  432. fib->vr = vr;
  433. fib->lpm_tree = lpm_tree;
  434. mlxsw_sp_lpm_tree_hold(lpm_tree);
  435. err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib, lpm_tree->id);
  436. if (err)
  437. goto err_lpm_tree_bind;
  438. return fib;
  439. err_lpm_tree_bind:
  440. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  441. err_rhashtable_init:
  442. kfree(fib);
  443. return ERR_PTR(err);
  444. }
  445. static void mlxsw_sp_fib_destroy(struct mlxsw_sp *mlxsw_sp,
  446. struct mlxsw_sp_fib *fib)
  447. {
  448. mlxsw_sp_vr_lpm_tree_unbind(mlxsw_sp, fib);
  449. mlxsw_sp_lpm_tree_put(mlxsw_sp, fib->lpm_tree);
  450. WARN_ON(!list_empty(&fib->node_list));
  451. rhashtable_destroy(&fib->ht);
  452. kfree(fib);
  453. }
  454. static struct mlxsw_sp_lpm_tree *
  455. mlxsw_sp_lpm_tree_find_unused(struct mlxsw_sp *mlxsw_sp)
  456. {
  457. static struct mlxsw_sp_lpm_tree *lpm_tree;
  458. int i;
  459. for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
  460. lpm_tree = &mlxsw_sp->router->lpm.trees[i];
  461. if (lpm_tree->ref_count == 0)
  462. return lpm_tree;
  463. }
  464. return NULL;
  465. }
  466. static int mlxsw_sp_lpm_tree_alloc(struct mlxsw_sp *mlxsw_sp,
  467. struct mlxsw_sp_lpm_tree *lpm_tree)
  468. {
  469. char ralta_pl[MLXSW_REG_RALTA_LEN];
  470. mlxsw_reg_ralta_pack(ralta_pl, true,
  471. (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
  472. lpm_tree->id);
  473. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
  474. }
  475. static void mlxsw_sp_lpm_tree_free(struct mlxsw_sp *mlxsw_sp,
  476. struct mlxsw_sp_lpm_tree *lpm_tree)
  477. {
  478. char ralta_pl[MLXSW_REG_RALTA_LEN];
  479. mlxsw_reg_ralta_pack(ralta_pl, false,
  480. (enum mlxsw_reg_ralxx_protocol) lpm_tree->proto,
  481. lpm_tree->id);
  482. mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
  483. }
  484. static int
  485. mlxsw_sp_lpm_tree_left_struct_set(struct mlxsw_sp *mlxsw_sp,
  486. struct mlxsw_sp_prefix_usage *prefix_usage,
  487. struct mlxsw_sp_lpm_tree *lpm_tree)
  488. {
  489. char ralst_pl[MLXSW_REG_RALST_LEN];
  490. u8 root_bin = 0;
  491. u8 prefix;
  492. u8 last_prefix = MLXSW_REG_RALST_BIN_NO_CHILD;
  493. mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage)
  494. root_bin = prefix;
  495. mlxsw_reg_ralst_pack(ralst_pl, root_bin, lpm_tree->id);
  496. mlxsw_sp_prefix_usage_for_each(prefix, prefix_usage) {
  497. if (prefix == 0)
  498. continue;
  499. mlxsw_reg_ralst_bin_pack(ralst_pl, prefix, last_prefix,
  500. MLXSW_REG_RALST_BIN_NO_CHILD);
  501. last_prefix = prefix;
  502. }
  503. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralst), ralst_pl);
  504. }
  505. static struct mlxsw_sp_lpm_tree *
  506. mlxsw_sp_lpm_tree_create(struct mlxsw_sp *mlxsw_sp,
  507. struct mlxsw_sp_prefix_usage *prefix_usage,
  508. enum mlxsw_sp_l3proto proto)
  509. {
  510. struct mlxsw_sp_lpm_tree *lpm_tree;
  511. int err;
  512. lpm_tree = mlxsw_sp_lpm_tree_find_unused(mlxsw_sp);
  513. if (!lpm_tree)
  514. return ERR_PTR(-EBUSY);
  515. lpm_tree->proto = proto;
  516. err = mlxsw_sp_lpm_tree_alloc(mlxsw_sp, lpm_tree);
  517. if (err)
  518. return ERR_PTR(err);
  519. err = mlxsw_sp_lpm_tree_left_struct_set(mlxsw_sp, prefix_usage,
  520. lpm_tree);
  521. if (err)
  522. goto err_left_struct_set;
  523. memcpy(&lpm_tree->prefix_usage, prefix_usage,
  524. sizeof(lpm_tree->prefix_usage));
  525. memset(&lpm_tree->prefix_ref_count, 0,
  526. sizeof(lpm_tree->prefix_ref_count));
  527. lpm_tree->ref_count = 1;
  528. return lpm_tree;
  529. err_left_struct_set:
  530. mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
  531. return ERR_PTR(err);
  532. }
  533. static void mlxsw_sp_lpm_tree_destroy(struct mlxsw_sp *mlxsw_sp,
  534. struct mlxsw_sp_lpm_tree *lpm_tree)
  535. {
  536. mlxsw_sp_lpm_tree_free(mlxsw_sp, lpm_tree);
  537. }
  538. static struct mlxsw_sp_lpm_tree *
  539. mlxsw_sp_lpm_tree_get(struct mlxsw_sp *mlxsw_sp,
  540. struct mlxsw_sp_prefix_usage *prefix_usage,
  541. enum mlxsw_sp_l3proto proto)
  542. {
  543. struct mlxsw_sp_lpm_tree *lpm_tree;
  544. int i;
  545. for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
  546. lpm_tree = &mlxsw_sp->router->lpm.trees[i];
  547. if (lpm_tree->ref_count != 0 &&
  548. lpm_tree->proto == proto &&
  549. mlxsw_sp_prefix_usage_eq(&lpm_tree->prefix_usage,
  550. prefix_usage)) {
  551. mlxsw_sp_lpm_tree_hold(lpm_tree);
  552. return lpm_tree;
  553. }
  554. }
  555. return mlxsw_sp_lpm_tree_create(mlxsw_sp, prefix_usage, proto);
  556. }
  557. static void mlxsw_sp_lpm_tree_hold(struct mlxsw_sp_lpm_tree *lpm_tree)
  558. {
  559. lpm_tree->ref_count++;
  560. }
  561. static void mlxsw_sp_lpm_tree_put(struct mlxsw_sp *mlxsw_sp,
  562. struct mlxsw_sp_lpm_tree *lpm_tree)
  563. {
  564. if (--lpm_tree->ref_count == 0)
  565. mlxsw_sp_lpm_tree_destroy(mlxsw_sp, lpm_tree);
  566. }
  567. #define MLXSW_SP_LPM_TREE_MIN 1 /* tree 0 is reserved */
  568. static int mlxsw_sp_lpm_init(struct mlxsw_sp *mlxsw_sp)
  569. {
  570. struct mlxsw_sp_prefix_usage req_prefix_usage = {{ 0 } };
  571. struct mlxsw_sp_lpm_tree *lpm_tree;
  572. u64 max_trees;
  573. int err, i;
  574. if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_LPM_TREES))
  575. return -EIO;
  576. max_trees = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_LPM_TREES);
  577. mlxsw_sp->router->lpm.tree_count = max_trees - MLXSW_SP_LPM_TREE_MIN;
  578. mlxsw_sp->router->lpm.trees = kcalloc(mlxsw_sp->router->lpm.tree_count,
  579. sizeof(struct mlxsw_sp_lpm_tree),
  580. GFP_KERNEL);
  581. if (!mlxsw_sp->router->lpm.trees)
  582. return -ENOMEM;
  583. for (i = 0; i < mlxsw_sp->router->lpm.tree_count; i++) {
  584. lpm_tree = &mlxsw_sp->router->lpm.trees[i];
  585. lpm_tree->id = i + MLXSW_SP_LPM_TREE_MIN;
  586. }
  587. lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
  588. MLXSW_SP_L3_PROTO_IPV4);
  589. if (IS_ERR(lpm_tree)) {
  590. err = PTR_ERR(lpm_tree);
  591. goto err_ipv4_tree_get;
  592. }
  593. mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4] = lpm_tree;
  594. lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
  595. MLXSW_SP_L3_PROTO_IPV6);
  596. if (IS_ERR(lpm_tree)) {
  597. err = PTR_ERR(lpm_tree);
  598. goto err_ipv6_tree_get;
  599. }
  600. mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV6] = lpm_tree;
  601. return 0;
  602. err_ipv6_tree_get:
  603. lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4];
  604. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  605. err_ipv4_tree_get:
  606. kfree(mlxsw_sp->router->lpm.trees);
  607. return err;
  608. }
  609. static void mlxsw_sp_lpm_fini(struct mlxsw_sp *mlxsw_sp)
  610. {
  611. struct mlxsw_sp_lpm_tree *lpm_tree;
  612. lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV6];
  613. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  614. lpm_tree = mlxsw_sp->router->lpm.proto_trees[MLXSW_SP_L3_PROTO_IPV4];
  615. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  616. kfree(mlxsw_sp->router->lpm.trees);
  617. }
  618. static bool mlxsw_sp_vr_is_used(const struct mlxsw_sp_vr *vr)
  619. {
  620. return !!vr->fib4 || !!vr->fib6 ||
  621. !!vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] ||
  622. !!vr->mr_table[MLXSW_SP_L3_PROTO_IPV6];
  623. }
  624. static struct mlxsw_sp_vr *mlxsw_sp_vr_find_unused(struct mlxsw_sp *mlxsw_sp)
  625. {
  626. struct mlxsw_sp_vr *vr;
  627. int i;
  628. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
  629. vr = &mlxsw_sp->router->vrs[i];
  630. if (!mlxsw_sp_vr_is_used(vr))
  631. return vr;
  632. }
  633. return NULL;
  634. }
  635. static int mlxsw_sp_vr_lpm_tree_bind(struct mlxsw_sp *mlxsw_sp,
  636. const struct mlxsw_sp_fib *fib, u8 tree_id)
  637. {
  638. char raltb_pl[MLXSW_REG_RALTB_LEN];
  639. mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
  640. (enum mlxsw_reg_ralxx_protocol) fib->proto,
  641. tree_id);
  642. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
  643. }
  644. static int mlxsw_sp_vr_lpm_tree_unbind(struct mlxsw_sp *mlxsw_sp,
  645. const struct mlxsw_sp_fib *fib)
  646. {
  647. char raltb_pl[MLXSW_REG_RALTB_LEN];
  648. /* Bind to tree 0 which is default */
  649. mlxsw_reg_raltb_pack(raltb_pl, fib->vr->id,
  650. (enum mlxsw_reg_ralxx_protocol) fib->proto, 0);
  651. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb), raltb_pl);
  652. }
  653. static u32 mlxsw_sp_fix_tb_id(u32 tb_id)
  654. {
  655. /* For our purpose, squash main, default and local tables into one */
  656. if (tb_id == RT_TABLE_LOCAL || tb_id == RT_TABLE_DEFAULT)
  657. tb_id = RT_TABLE_MAIN;
  658. return tb_id;
  659. }
  660. static struct mlxsw_sp_vr *mlxsw_sp_vr_find(struct mlxsw_sp *mlxsw_sp,
  661. u32 tb_id)
  662. {
  663. struct mlxsw_sp_vr *vr;
  664. int i;
  665. tb_id = mlxsw_sp_fix_tb_id(tb_id);
  666. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
  667. vr = &mlxsw_sp->router->vrs[i];
  668. if (mlxsw_sp_vr_is_used(vr) && vr->tb_id == tb_id)
  669. return vr;
  670. }
  671. return NULL;
  672. }
  673. static struct mlxsw_sp_fib *mlxsw_sp_vr_fib(const struct mlxsw_sp_vr *vr,
  674. enum mlxsw_sp_l3proto proto)
  675. {
  676. switch (proto) {
  677. case MLXSW_SP_L3_PROTO_IPV4:
  678. return vr->fib4;
  679. case MLXSW_SP_L3_PROTO_IPV6:
  680. return vr->fib6;
  681. }
  682. return NULL;
  683. }
  684. static struct mlxsw_sp_vr *mlxsw_sp_vr_create(struct mlxsw_sp *mlxsw_sp,
  685. u32 tb_id,
  686. struct netlink_ext_ack *extack)
  687. {
  688. struct mlxsw_sp_mr_table *mr4_table, *mr6_table;
  689. struct mlxsw_sp_fib *fib4;
  690. struct mlxsw_sp_fib *fib6;
  691. struct mlxsw_sp_vr *vr;
  692. int err;
  693. vr = mlxsw_sp_vr_find_unused(mlxsw_sp);
  694. if (!vr) {
  695. NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported virtual routers");
  696. return ERR_PTR(-EBUSY);
  697. }
  698. fib4 = mlxsw_sp_fib_create(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV4);
  699. if (IS_ERR(fib4))
  700. return ERR_CAST(fib4);
  701. fib6 = mlxsw_sp_fib_create(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV6);
  702. if (IS_ERR(fib6)) {
  703. err = PTR_ERR(fib6);
  704. goto err_fib6_create;
  705. }
  706. mr4_table = mlxsw_sp_mr_table_create(mlxsw_sp, vr->id,
  707. MLXSW_SP_L3_PROTO_IPV4);
  708. if (IS_ERR(mr4_table)) {
  709. err = PTR_ERR(mr4_table);
  710. goto err_mr4_table_create;
  711. }
  712. mr6_table = mlxsw_sp_mr_table_create(mlxsw_sp, vr->id,
  713. MLXSW_SP_L3_PROTO_IPV6);
  714. if (IS_ERR(mr6_table)) {
  715. err = PTR_ERR(mr6_table);
  716. goto err_mr6_table_create;
  717. }
  718. vr->fib4 = fib4;
  719. vr->fib6 = fib6;
  720. vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] = mr4_table;
  721. vr->mr_table[MLXSW_SP_L3_PROTO_IPV6] = mr6_table;
  722. vr->tb_id = tb_id;
  723. return vr;
  724. err_mr6_table_create:
  725. mlxsw_sp_mr_table_destroy(mr4_table);
  726. err_mr4_table_create:
  727. mlxsw_sp_fib_destroy(mlxsw_sp, fib6);
  728. err_fib6_create:
  729. mlxsw_sp_fib_destroy(mlxsw_sp, fib4);
  730. return ERR_PTR(err);
  731. }
  732. static void mlxsw_sp_vr_destroy(struct mlxsw_sp *mlxsw_sp,
  733. struct mlxsw_sp_vr *vr)
  734. {
  735. mlxsw_sp_mr_table_destroy(vr->mr_table[MLXSW_SP_L3_PROTO_IPV6]);
  736. vr->mr_table[MLXSW_SP_L3_PROTO_IPV6] = NULL;
  737. mlxsw_sp_mr_table_destroy(vr->mr_table[MLXSW_SP_L3_PROTO_IPV4]);
  738. vr->mr_table[MLXSW_SP_L3_PROTO_IPV4] = NULL;
  739. mlxsw_sp_fib_destroy(mlxsw_sp, vr->fib6);
  740. vr->fib6 = NULL;
  741. mlxsw_sp_fib_destroy(mlxsw_sp, vr->fib4);
  742. vr->fib4 = NULL;
  743. }
  744. static struct mlxsw_sp_vr *mlxsw_sp_vr_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id,
  745. struct netlink_ext_ack *extack)
  746. {
  747. struct mlxsw_sp_vr *vr;
  748. tb_id = mlxsw_sp_fix_tb_id(tb_id);
  749. vr = mlxsw_sp_vr_find(mlxsw_sp, tb_id);
  750. if (!vr)
  751. vr = mlxsw_sp_vr_create(mlxsw_sp, tb_id, extack);
  752. return vr;
  753. }
  754. static void mlxsw_sp_vr_put(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_vr *vr)
  755. {
  756. if (!vr->rif_count && list_empty(&vr->fib4->node_list) &&
  757. list_empty(&vr->fib6->node_list) &&
  758. mlxsw_sp_mr_table_empty(vr->mr_table[MLXSW_SP_L3_PROTO_IPV4]) &&
  759. mlxsw_sp_mr_table_empty(vr->mr_table[MLXSW_SP_L3_PROTO_IPV6]))
  760. mlxsw_sp_vr_destroy(mlxsw_sp, vr);
  761. }
  762. static bool
  763. mlxsw_sp_vr_lpm_tree_should_replace(struct mlxsw_sp_vr *vr,
  764. enum mlxsw_sp_l3proto proto, u8 tree_id)
  765. {
  766. struct mlxsw_sp_fib *fib = mlxsw_sp_vr_fib(vr, proto);
  767. if (!mlxsw_sp_vr_is_used(vr))
  768. return false;
  769. if (fib->lpm_tree->id == tree_id)
  770. return true;
  771. return false;
  772. }
  773. static int mlxsw_sp_vr_lpm_tree_replace(struct mlxsw_sp *mlxsw_sp,
  774. struct mlxsw_sp_fib *fib,
  775. struct mlxsw_sp_lpm_tree *new_tree)
  776. {
  777. struct mlxsw_sp_lpm_tree *old_tree = fib->lpm_tree;
  778. int err;
  779. fib->lpm_tree = new_tree;
  780. mlxsw_sp_lpm_tree_hold(new_tree);
  781. err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, fib, new_tree->id);
  782. if (err)
  783. goto err_tree_bind;
  784. mlxsw_sp_lpm_tree_put(mlxsw_sp, old_tree);
  785. return 0;
  786. err_tree_bind:
  787. mlxsw_sp_lpm_tree_put(mlxsw_sp, new_tree);
  788. fib->lpm_tree = old_tree;
  789. return err;
  790. }
  791. static int mlxsw_sp_vrs_lpm_tree_replace(struct mlxsw_sp *mlxsw_sp,
  792. struct mlxsw_sp_fib *fib,
  793. struct mlxsw_sp_lpm_tree *new_tree)
  794. {
  795. enum mlxsw_sp_l3proto proto = fib->proto;
  796. struct mlxsw_sp_lpm_tree *old_tree;
  797. u8 old_id, new_id = new_tree->id;
  798. struct mlxsw_sp_vr *vr;
  799. int i, err;
  800. old_tree = mlxsw_sp->router->lpm.proto_trees[proto];
  801. old_id = old_tree->id;
  802. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
  803. vr = &mlxsw_sp->router->vrs[i];
  804. if (!mlxsw_sp_vr_lpm_tree_should_replace(vr, proto, old_id))
  805. continue;
  806. err = mlxsw_sp_vr_lpm_tree_replace(mlxsw_sp,
  807. mlxsw_sp_vr_fib(vr, proto),
  808. new_tree);
  809. if (err)
  810. goto err_tree_replace;
  811. }
  812. memcpy(new_tree->prefix_ref_count, old_tree->prefix_ref_count,
  813. sizeof(new_tree->prefix_ref_count));
  814. mlxsw_sp->router->lpm.proto_trees[proto] = new_tree;
  815. mlxsw_sp_lpm_tree_put(mlxsw_sp, old_tree);
  816. return 0;
  817. err_tree_replace:
  818. for (i--; i >= 0; i--) {
  819. if (!mlxsw_sp_vr_lpm_tree_should_replace(vr, proto, new_id))
  820. continue;
  821. mlxsw_sp_vr_lpm_tree_replace(mlxsw_sp,
  822. mlxsw_sp_vr_fib(vr, proto),
  823. old_tree);
  824. }
  825. return err;
  826. }
  827. static int mlxsw_sp_vrs_init(struct mlxsw_sp *mlxsw_sp)
  828. {
  829. struct mlxsw_sp_vr *vr;
  830. u64 max_vrs;
  831. int i;
  832. if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_VRS))
  833. return -EIO;
  834. max_vrs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS);
  835. mlxsw_sp->router->vrs = kcalloc(max_vrs, sizeof(struct mlxsw_sp_vr),
  836. GFP_KERNEL);
  837. if (!mlxsw_sp->router->vrs)
  838. return -ENOMEM;
  839. for (i = 0; i < max_vrs; i++) {
  840. vr = &mlxsw_sp->router->vrs[i];
  841. vr->id = i;
  842. }
  843. return 0;
  844. }
  845. static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp);
  846. static void mlxsw_sp_vrs_fini(struct mlxsw_sp *mlxsw_sp)
  847. {
  848. /* At this stage we're guaranteed not to have new incoming
  849. * FIB notifications and the work queue is free from FIBs
  850. * sitting on top of mlxsw netdevs. However, we can still
  851. * have other FIBs queued. Flush the queue before flushing
  852. * the device's tables. No need for locks, as we're the only
  853. * writer.
  854. */
  855. mlxsw_core_flush_owq();
  856. mlxsw_sp_router_fib_flush(mlxsw_sp);
  857. kfree(mlxsw_sp->router->vrs);
  858. }
  859. static struct net_device *
  860. __mlxsw_sp_ipip_netdev_ul_dev_get(const struct net_device *ol_dev)
  861. {
  862. struct ip_tunnel *tun = netdev_priv(ol_dev);
  863. struct net *net = dev_net(ol_dev);
  864. return __dev_get_by_index(net, tun->parms.link);
  865. }
  866. u32 mlxsw_sp_ipip_dev_ul_tb_id(const struct net_device *ol_dev)
  867. {
  868. struct net_device *d = __mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
  869. if (d)
  870. return l3mdev_fib_table(d) ? : RT_TABLE_MAIN;
  871. else
  872. return l3mdev_fib_table(ol_dev) ? : RT_TABLE_MAIN;
  873. }
  874. static struct mlxsw_sp_rif *
  875. mlxsw_sp_rif_create(struct mlxsw_sp *mlxsw_sp,
  876. const struct mlxsw_sp_rif_params *params,
  877. struct netlink_ext_ack *extack);
  878. static struct mlxsw_sp_rif_ipip_lb *
  879. mlxsw_sp_ipip_ol_ipip_lb_create(struct mlxsw_sp *mlxsw_sp,
  880. enum mlxsw_sp_ipip_type ipipt,
  881. struct net_device *ol_dev,
  882. struct netlink_ext_ack *extack)
  883. {
  884. struct mlxsw_sp_rif_params_ipip_lb lb_params;
  885. const struct mlxsw_sp_ipip_ops *ipip_ops;
  886. struct mlxsw_sp_rif *rif;
  887. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipipt];
  888. lb_params = (struct mlxsw_sp_rif_params_ipip_lb) {
  889. .common.dev = ol_dev,
  890. .common.lag = false,
  891. .lb_config = ipip_ops->ol_loopback_config(mlxsw_sp, ol_dev),
  892. };
  893. rif = mlxsw_sp_rif_create(mlxsw_sp, &lb_params.common, extack);
  894. if (IS_ERR(rif))
  895. return ERR_CAST(rif);
  896. return container_of(rif, struct mlxsw_sp_rif_ipip_lb, common);
  897. }
  898. static struct mlxsw_sp_ipip_entry *
  899. mlxsw_sp_ipip_entry_alloc(struct mlxsw_sp *mlxsw_sp,
  900. enum mlxsw_sp_ipip_type ipipt,
  901. struct net_device *ol_dev)
  902. {
  903. const struct mlxsw_sp_ipip_ops *ipip_ops;
  904. struct mlxsw_sp_ipip_entry *ipip_entry;
  905. struct mlxsw_sp_ipip_entry *ret = NULL;
  906. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipipt];
  907. ipip_entry = kzalloc(sizeof(*ipip_entry), GFP_KERNEL);
  908. if (!ipip_entry)
  909. return ERR_PTR(-ENOMEM);
  910. ipip_entry->ol_lb = mlxsw_sp_ipip_ol_ipip_lb_create(mlxsw_sp, ipipt,
  911. ol_dev, NULL);
  912. if (IS_ERR(ipip_entry->ol_lb)) {
  913. ret = ERR_CAST(ipip_entry->ol_lb);
  914. goto err_ol_ipip_lb_create;
  915. }
  916. ipip_entry->ipipt = ipipt;
  917. ipip_entry->ol_dev = ol_dev;
  918. switch (ipip_ops->ul_proto) {
  919. case MLXSW_SP_L3_PROTO_IPV4:
  920. ipip_entry->parms4 = mlxsw_sp_ipip_netdev_parms4(ol_dev);
  921. break;
  922. case MLXSW_SP_L3_PROTO_IPV6:
  923. WARN_ON(1);
  924. break;
  925. }
  926. return ipip_entry;
  927. err_ol_ipip_lb_create:
  928. kfree(ipip_entry);
  929. return ret;
  930. }
  931. static void
  932. mlxsw_sp_ipip_entry_dealloc(struct mlxsw_sp_ipip_entry *ipip_entry)
  933. {
  934. mlxsw_sp_rif_destroy(&ipip_entry->ol_lb->common);
  935. kfree(ipip_entry);
  936. }
  937. static bool
  938. mlxsw_sp_ipip_entry_saddr_matches(struct mlxsw_sp *mlxsw_sp,
  939. const enum mlxsw_sp_l3proto ul_proto,
  940. union mlxsw_sp_l3addr saddr,
  941. u32 ul_tb_id,
  942. struct mlxsw_sp_ipip_entry *ipip_entry)
  943. {
  944. u32 tun_ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ipip_entry->ol_dev);
  945. enum mlxsw_sp_ipip_type ipipt = ipip_entry->ipipt;
  946. union mlxsw_sp_l3addr tun_saddr;
  947. if (mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto != ul_proto)
  948. return false;
  949. tun_saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ipip_entry->ol_dev);
  950. return tun_ul_tb_id == ul_tb_id &&
  951. mlxsw_sp_l3addr_eq(&tun_saddr, &saddr);
  952. }
  953. static int
  954. mlxsw_sp_fib_entry_decap_init(struct mlxsw_sp *mlxsw_sp,
  955. struct mlxsw_sp_fib_entry *fib_entry,
  956. struct mlxsw_sp_ipip_entry *ipip_entry)
  957. {
  958. u32 tunnel_index;
  959. int err;
  960. err = mlxsw_sp_kvdl_alloc(mlxsw_sp, 1, &tunnel_index);
  961. if (err)
  962. return err;
  963. ipip_entry->decap_fib_entry = fib_entry;
  964. fib_entry->decap.ipip_entry = ipip_entry;
  965. fib_entry->decap.tunnel_index = tunnel_index;
  966. return 0;
  967. }
  968. static void mlxsw_sp_fib_entry_decap_fini(struct mlxsw_sp *mlxsw_sp,
  969. struct mlxsw_sp_fib_entry *fib_entry)
  970. {
  971. /* Unlink this node from the IPIP entry that it's the decap entry of. */
  972. fib_entry->decap.ipip_entry->decap_fib_entry = NULL;
  973. fib_entry->decap.ipip_entry = NULL;
  974. mlxsw_sp_kvdl_free(mlxsw_sp, fib_entry->decap.tunnel_index);
  975. }
  976. static struct mlxsw_sp_fib_node *
  977. mlxsw_sp_fib_node_lookup(struct mlxsw_sp_fib *fib, const void *addr,
  978. size_t addr_len, unsigned char prefix_len);
  979. static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
  980. struct mlxsw_sp_fib_entry *fib_entry);
  981. static void
  982. mlxsw_sp_ipip_entry_demote_decap(struct mlxsw_sp *mlxsw_sp,
  983. struct mlxsw_sp_ipip_entry *ipip_entry)
  984. {
  985. struct mlxsw_sp_fib_entry *fib_entry = ipip_entry->decap_fib_entry;
  986. mlxsw_sp_fib_entry_decap_fini(mlxsw_sp, fib_entry);
  987. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
  988. mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
  989. }
  990. static void
  991. mlxsw_sp_ipip_entry_promote_decap(struct mlxsw_sp *mlxsw_sp,
  992. struct mlxsw_sp_ipip_entry *ipip_entry,
  993. struct mlxsw_sp_fib_entry *decap_fib_entry)
  994. {
  995. if (mlxsw_sp_fib_entry_decap_init(mlxsw_sp, decap_fib_entry,
  996. ipip_entry))
  997. return;
  998. decap_fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP;
  999. if (mlxsw_sp_fib_entry_update(mlxsw_sp, decap_fib_entry))
  1000. mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
  1001. }
  1002. /* Given an IPIP entry, find the corresponding decap route. */
  1003. static struct mlxsw_sp_fib_entry *
  1004. mlxsw_sp_ipip_entry_find_decap(struct mlxsw_sp *mlxsw_sp,
  1005. struct mlxsw_sp_ipip_entry *ipip_entry)
  1006. {
  1007. static struct mlxsw_sp_fib_node *fib_node;
  1008. const struct mlxsw_sp_ipip_ops *ipip_ops;
  1009. struct mlxsw_sp_fib_entry *fib_entry;
  1010. unsigned char saddr_prefix_len;
  1011. union mlxsw_sp_l3addr saddr;
  1012. struct mlxsw_sp_fib *ul_fib;
  1013. struct mlxsw_sp_vr *ul_vr;
  1014. const void *saddrp;
  1015. size_t saddr_len;
  1016. u32 ul_tb_id;
  1017. u32 saddr4;
  1018. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
  1019. ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ipip_entry->ol_dev);
  1020. ul_vr = mlxsw_sp_vr_find(mlxsw_sp, ul_tb_id);
  1021. if (!ul_vr)
  1022. return NULL;
  1023. ul_fib = mlxsw_sp_vr_fib(ul_vr, ipip_ops->ul_proto);
  1024. saddr = mlxsw_sp_ipip_netdev_saddr(ipip_ops->ul_proto,
  1025. ipip_entry->ol_dev);
  1026. switch (ipip_ops->ul_proto) {
  1027. case MLXSW_SP_L3_PROTO_IPV4:
  1028. saddr4 = be32_to_cpu(saddr.addr4);
  1029. saddrp = &saddr4;
  1030. saddr_len = 4;
  1031. saddr_prefix_len = 32;
  1032. break;
  1033. case MLXSW_SP_L3_PROTO_IPV6:
  1034. WARN_ON(1);
  1035. return NULL;
  1036. }
  1037. fib_node = mlxsw_sp_fib_node_lookup(ul_fib, saddrp, saddr_len,
  1038. saddr_prefix_len);
  1039. if (!fib_node || list_empty(&fib_node->entry_list))
  1040. return NULL;
  1041. fib_entry = list_first_entry(&fib_node->entry_list,
  1042. struct mlxsw_sp_fib_entry, list);
  1043. if (fib_entry->type != MLXSW_SP_FIB_ENTRY_TYPE_TRAP)
  1044. return NULL;
  1045. return fib_entry;
  1046. }
  1047. static struct mlxsw_sp_ipip_entry *
  1048. mlxsw_sp_ipip_entry_create(struct mlxsw_sp *mlxsw_sp,
  1049. enum mlxsw_sp_ipip_type ipipt,
  1050. struct net_device *ol_dev)
  1051. {
  1052. struct mlxsw_sp_ipip_entry *ipip_entry;
  1053. ipip_entry = mlxsw_sp_ipip_entry_alloc(mlxsw_sp, ipipt, ol_dev);
  1054. if (IS_ERR(ipip_entry))
  1055. return ipip_entry;
  1056. list_add_tail(&ipip_entry->ipip_list_node,
  1057. &mlxsw_sp->router->ipip_list);
  1058. return ipip_entry;
  1059. }
  1060. static void
  1061. mlxsw_sp_ipip_entry_destroy(struct mlxsw_sp *mlxsw_sp,
  1062. struct mlxsw_sp_ipip_entry *ipip_entry)
  1063. {
  1064. list_del(&ipip_entry->ipip_list_node);
  1065. mlxsw_sp_ipip_entry_dealloc(ipip_entry);
  1066. }
  1067. static bool
  1068. mlxsw_sp_ipip_entry_matches_decap(struct mlxsw_sp *mlxsw_sp,
  1069. const struct net_device *ul_dev,
  1070. enum mlxsw_sp_l3proto ul_proto,
  1071. union mlxsw_sp_l3addr ul_dip,
  1072. struct mlxsw_sp_ipip_entry *ipip_entry)
  1073. {
  1074. u32 ul_tb_id = l3mdev_fib_table(ul_dev) ? : RT_TABLE_MAIN;
  1075. enum mlxsw_sp_ipip_type ipipt = ipip_entry->ipipt;
  1076. struct net_device *ipip_ul_dev;
  1077. if (mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto != ul_proto)
  1078. return false;
  1079. ipip_ul_dev = __mlxsw_sp_ipip_netdev_ul_dev_get(ipip_entry->ol_dev);
  1080. return mlxsw_sp_ipip_entry_saddr_matches(mlxsw_sp, ul_proto, ul_dip,
  1081. ul_tb_id, ipip_entry) &&
  1082. (!ipip_ul_dev || ipip_ul_dev == ul_dev);
  1083. }
  1084. /* Given decap parameters, find the corresponding IPIP entry. */
  1085. static struct mlxsw_sp_ipip_entry *
  1086. mlxsw_sp_ipip_entry_find_by_decap(struct mlxsw_sp *mlxsw_sp,
  1087. const struct net_device *ul_dev,
  1088. enum mlxsw_sp_l3proto ul_proto,
  1089. union mlxsw_sp_l3addr ul_dip)
  1090. {
  1091. struct mlxsw_sp_ipip_entry *ipip_entry;
  1092. list_for_each_entry(ipip_entry, &mlxsw_sp->router->ipip_list,
  1093. ipip_list_node)
  1094. if (mlxsw_sp_ipip_entry_matches_decap(mlxsw_sp, ul_dev,
  1095. ul_proto, ul_dip,
  1096. ipip_entry))
  1097. return ipip_entry;
  1098. return NULL;
  1099. }
  1100. static bool mlxsw_sp_netdev_ipip_type(const struct mlxsw_sp *mlxsw_sp,
  1101. const struct net_device *dev,
  1102. enum mlxsw_sp_ipip_type *p_type)
  1103. {
  1104. struct mlxsw_sp_router *router = mlxsw_sp->router;
  1105. const struct mlxsw_sp_ipip_ops *ipip_ops;
  1106. enum mlxsw_sp_ipip_type ipipt;
  1107. for (ipipt = 0; ipipt < MLXSW_SP_IPIP_TYPE_MAX; ++ipipt) {
  1108. ipip_ops = router->ipip_ops_arr[ipipt];
  1109. if (dev->type == ipip_ops->dev_type) {
  1110. if (p_type)
  1111. *p_type = ipipt;
  1112. return true;
  1113. }
  1114. }
  1115. return false;
  1116. }
  1117. bool mlxsw_sp_netdev_is_ipip_ol(const struct mlxsw_sp *mlxsw_sp,
  1118. const struct net_device *dev)
  1119. {
  1120. return mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, NULL);
  1121. }
  1122. static struct mlxsw_sp_ipip_entry *
  1123. mlxsw_sp_ipip_entry_find_by_ol_dev(struct mlxsw_sp *mlxsw_sp,
  1124. const struct net_device *ol_dev)
  1125. {
  1126. struct mlxsw_sp_ipip_entry *ipip_entry;
  1127. list_for_each_entry(ipip_entry, &mlxsw_sp->router->ipip_list,
  1128. ipip_list_node)
  1129. if (ipip_entry->ol_dev == ol_dev)
  1130. return ipip_entry;
  1131. return NULL;
  1132. }
  1133. static struct mlxsw_sp_ipip_entry *
  1134. mlxsw_sp_ipip_entry_find_by_ul_dev(const struct mlxsw_sp *mlxsw_sp,
  1135. const struct net_device *ul_dev,
  1136. struct mlxsw_sp_ipip_entry *start)
  1137. {
  1138. struct mlxsw_sp_ipip_entry *ipip_entry;
  1139. ipip_entry = list_prepare_entry(start, &mlxsw_sp->router->ipip_list,
  1140. ipip_list_node);
  1141. list_for_each_entry_continue(ipip_entry, &mlxsw_sp->router->ipip_list,
  1142. ipip_list_node) {
  1143. struct net_device *ipip_ul_dev =
  1144. __mlxsw_sp_ipip_netdev_ul_dev_get(ipip_entry->ol_dev);
  1145. if (ipip_ul_dev == ul_dev)
  1146. return ipip_entry;
  1147. }
  1148. return NULL;
  1149. }
  1150. bool mlxsw_sp_netdev_is_ipip_ul(const struct mlxsw_sp *mlxsw_sp,
  1151. const struct net_device *dev)
  1152. {
  1153. return mlxsw_sp_ipip_entry_find_by_ul_dev(mlxsw_sp, dev, NULL);
  1154. }
  1155. static bool mlxsw_sp_netdevice_ipip_can_offload(struct mlxsw_sp *mlxsw_sp,
  1156. const struct net_device *ol_dev,
  1157. enum mlxsw_sp_ipip_type ipipt)
  1158. {
  1159. const struct mlxsw_sp_ipip_ops *ops
  1160. = mlxsw_sp->router->ipip_ops_arr[ipipt];
  1161. /* For deciding whether decap should be offloaded, we don't care about
  1162. * overlay protocol, so ask whether either one is supported.
  1163. */
  1164. return ops->can_offload(mlxsw_sp, ol_dev, MLXSW_SP_L3_PROTO_IPV4) ||
  1165. ops->can_offload(mlxsw_sp, ol_dev, MLXSW_SP_L3_PROTO_IPV6);
  1166. }
  1167. static int mlxsw_sp_netdevice_ipip_ol_reg_event(struct mlxsw_sp *mlxsw_sp,
  1168. struct net_device *ol_dev)
  1169. {
  1170. struct mlxsw_sp_ipip_entry *ipip_entry;
  1171. enum mlxsw_sp_l3proto ul_proto;
  1172. enum mlxsw_sp_ipip_type ipipt;
  1173. union mlxsw_sp_l3addr saddr;
  1174. u32 ul_tb_id;
  1175. mlxsw_sp_netdev_ipip_type(mlxsw_sp, ol_dev, &ipipt);
  1176. if (mlxsw_sp_netdevice_ipip_can_offload(mlxsw_sp, ol_dev, ipipt)) {
  1177. ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ol_dev);
  1178. ul_proto = mlxsw_sp->router->ipip_ops_arr[ipipt]->ul_proto;
  1179. saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ol_dev);
  1180. if (!mlxsw_sp_ipip_demote_tunnel_by_saddr(mlxsw_sp, ul_proto,
  1181. saddr, ul_tb_id,
  1182. NULL)) {
  1183. ipip_entry = mlxsw_sp_ipip_entry_create(mlxsw_sp, ipipt,
  1184. ol_dev);
  1185. if (IS_ERR(ipip_entry))
  1186. return PTR_ERR(ipip_entry);
  1187. }
  1188. }
  1189. return 0;
  1190. }
  1191. static void mlxsw_sp_netdevice_ipip_ol_unreg_event(struct mlxsw_sp *mlxsw_sp,
  1192. struct net_device *ol_dev)
  1193. {
  1194. struct mlxsw_sp_ipip_entry *ipip_entry;
  1195. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1196. if (ipip_entry)
  1197. mlxsw_sp_ipip_entry_destroy(mlxsw_sp, ipip_entry);
  1198. }
  1199. static void
  1200. mlxsw_sp_ipip_entry_ol_up_event(struct mlxsw_sp *mlxsw_sp,
  1201. struct mlxsw_sp_ipip_entry *ipip_entry)
  1202. {
  1203. struct mlxsw_sp_fib_entry *decap_fib_entry;
  1204. decap_fib_entry = mlxsw_sp_ipip_entry_find_decap(mlxsw_sp, ipip_entry);
  1205. if (decap_fib_entry)
  1206. mlxsw_sp_ipip_entry_promote_decap(mlxsw_sp, ipip_entry,
  1207. decap_fib_entry);
  1208. }
  1209. static int
  1210. mlxsw_sp_rif_ipip_lb_op(struct mlxsw_sp_rif_ipip_lb *lb_rif,
  1211. struct mlxsw_sp_vr *ul_vr, bool enable)
  1212. {
  1213. struct mlxsw_sp_rif_ipip_lb_config lb_cf = lb_rif->lb_config;
  1214. struct mlxsw_sp_rif *rif = &lb_rif->common;
  1215. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  1216. char ritr_pl[MLXSW_REG_RITR_LEN];
  1217. u32 saddr4;
  1218. switch (lb_cf.ul_protocol) {
  1219. case MLXSW_SP_L3_PROTO_IPV4:
  1220. saddr4 = be32_to_cpu(lb_cf.saddr.addr4);
  1221. mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_LOOPBACK_IF,
  1222. rif->rif_index, rif->vr_id, rif->dev->mtu);
  1223. mlxsw_reg_ritr_loopback_ipip4_pack(ritr_pl, lb_cf.lb_ipipt,
  1224. MLXSW_REG_RITR_LOOPBACK_IPIP_OPTIONS_GRE_KEY_PRESET,
  1225. ul_vr->id, saddr4, lb_cf.okey);
  1226. break;
  1227. case MLXSW_SP_L3_PROTO_IPV6:
  1228. return -EAFNOSUPPORT;
  1229. }
  1230. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  1231. }
  1232. static int mlxsw_sp_netdevice_ipip_ol_update_mtu(struct mlxsw_sp *mlxsw_sp,
  1233. struct net_device *ol_dev)
  1234. {
  1235. struct mlxsw_sp_ipip_entry *ipip_entry;
  1236. struct mlxsw_sp_rif_ipip_lb *lb_rif;
  1237. struct mlxsw_sp_vr *ul_vr;
  1238. int err = 0;
  1239. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1240. if (ipip_entry) {
  1241. lb_rif = ipip_entry->ol_lb;
  1242. ul_vr = &mlxsw_sp->router->vrs[lb_rif->ul_vr_id];
  1243. err = mlxsw_sp_rif_ipip_lb_op(lb_rif, ul_vr, true);
  1244. if (err)
  1245. goto out;
  1246. lb_rif->common.mtu = ol_dev->mtu;
  1247. }
  1248. out:
  1249. return err;
  1250. }
  1251. static void mlxsw_sp_netdevice_ipip_ol_up_event(struct mlxsw_sp *mlxsw_sp,
  1252. struct net_device *ol_dev)
  1253. {
  1254. struct mlxsw_sp_ipip_entry *ipip_entry;
  1255. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1256. if (ipip_entry)
  1257. mlxsw_sp_ipip_entry_ol_up_event(mlxsw_sp, ipip_entry);
  1258. }
  1259. static void
  1260. mlxsw_sp_ipip_entry_ol_down_event(struct mlxsw_sp *mlxsw_sp,
  1261. struct mlxsw_sp_ipip_entry *ipip_entry)
  1262. {
  1263. if (ipip_entry->decap_fib_entry)
  1264. mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
  1265. }
  1266. static void mlxsw_sp_netdevice_ipip_ol_down_event(struct mlxsw_sp *mlxsw_sp,
  1267. struct net_device *ol_dev)
  1268. {
  1269. struct mlxsw_sp_ipip_entry *ipip_entry;
  1270. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1271. if (ipip_entry)
  1272. mlxsw_sp_ipip_entry_ol_down_event(mlxsw_sp, ipip_entry);
  1273. }
  1274. static void mlxsw_sp_nexthop_rif_migrate(struct mlxsw_sp *mlxsw_sp,
  1275. struct mlxsw_sp_rif *old_rif,
  1276. struct mlxsw_sp_rif *new_rif);
  1277. static int
  1278. mlxsw_sp_ipip_entry_ol_lb_update(struct mlxsw_sp *mlxsw_sp,
  1279. struct mlxsw_sp_ipip_entry *ipip_entry,
  1280. bool keep_encap,
  1281. struct netlink_ext_ack *extack)
  1282. {
  1283. struct mlxsw_sp_rif_ipip_lb *old_lb_rif = ipip_entry->ol_lb;
  1284. struct mlxsw_sp_rif_ipip_lb *new_lb_rif;
  1285. new_lb_rif = mlxsw_sp_ipip_ol_ipip_lb_create(mlxsw_sp,
  1286. ipip_entry->ipipt,
  1287. ipip_entry->ol_dev,
  1288. extack);
  1289. if (IS_ERR(new_lb_rif))
  1290. return PTR_ERR(new_lb_rif);
  1291. ipip_entry->ol_lb = new_lb_rif;
  1292. if (keep_encap)
  1293. mlxsw_sp_nexthop_rif_migrate(mlxsw_sp, &old_lb_rif->common,
  1294. &new_lb_rif->common);
  1295. mlxsw_sp_rif_destroy(&old_lb_rif->common);
  1296. return 0;
  1297. }
  1298. static void mlxsw_sp_nexthop_rif_update(struct mlxsw_sp *mlxsw_sp,
  1299. struct mlxsw_sp_rif *rif);
  1300. /**
  1301. * Update the offload related to an IPIP entry. This always updates decap, and
  1302. * in addition to that it also:
  1303. * @recreate_loopback: recreates the associated loopback RIF
  1304. * @keep_encap: updates next hops that use the tunnel netdevice. This is only
  1305. * relevant when recreate_loopback is true.
  1306. * @update_nexthops: updates next hops, keeping the current loopback RIF. This
  1307. * is only relevant when recreate_loopback is false.
  1308. */
  1309. int __mlxsw_sp_ipip_entry_update_tunnel(struct mlxsw_sp *mlxsw_sp,
  1310. struct mlxsw_sp_ipip_entry *ipip_entry,
  1311. bool recreate_loopback,
  1312. bool keep_encap,
  1313. bool update_nexthops,
  1314. struct netlink_ext_ack *extack)
  1315. {
  1316. int err;
  1317. /* RIFs can't be edited, so to update loopback, we need to destroy and
  1318. * recreate it. That creates a window of opportunity where RALUE and
  1319. * RATR registers end up referencing a RIF that's already gone. RATRs
  1320. * are handled in mlxsw_sp_ipip_entry_ol_lb_update(), and to take care
  1321. * of RALUE, demote the decap route back.
  1322. */
  1323. if (ipip_entry->decap_fib_entry)
  1324. mlxsw_sp_ipip_entry_demote_decap(mlxsw_sp, ipip_entry);
  1325. if (recreate_loopback) {
  1326. err = mlxsw_sp_ipip_entry_ol_lb_update(mlxsw_sp, ipip_entry,
  1327. keep_encap, extack);
  1328. if (err)
  1329. return err;
  1330. } else if (update_nexthops) {
  1331. mlxsw_sp_nexthop_rif_update(mlxsw_sp,
  1332. &ipip_entry->ol_lb->common);
  1333. }
  1334. if (ipip_entry->ol_dev->flags & IFF_UP)
  1335. mlxsw_sp_ipip_entry_ol_up_event(mlxsw_sp, ipip_entry);
  1336. return 0;
  1337. }
  1338. static int mlxsw_sp_netdevice_ipip_ol_vrf_event(struct mlxsw_sp *mlxsw_sp,
  1339. struct net_device *ol_dev,
  1340. struct netlink_ext_ack *extack)
  1341. {
  1342. struct mlxsw_sp_ipip_entry *ipip_entry =
  1343. mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1344. enum mlxsw_sp_l3proto ul_proto;
  1345. union mlxsw_sp_l3addr saddr;
  1346. u32 ul_tb_id;
  1347. if (!ipip_entry)
  1348. return 0;
  1349. /* For flat configuration cases, moving overlay to a different VRF might
  1350. * cause local address conflict, and the conflicting tunnels need to be
  1351. * demoted.
  1352. */
  1353. ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(ol_dev);
  1354. ul_proto = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt]->ul_proto;
  1355. saddr = mlxsw_sp_ipip_netdev_saddr(ul_proto, ol_dev);
  1356. if (mlxsw_sp_ipip_demote_tunnel_by_saddr(mlxsw_sp, ul_proto,
  1357. saddr, ul_tb_id,
  1358. ipip_entry)) {
  1359. mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
  1360. return 0;
  1361. }
  1362. return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
  1363. true, false, false, extack);
  1364. }
  1365. static int
  1366. mlxsw_sp_netdevice_ipip_ul_vrf_event(struct mlxsw_sp *mlxsw_sp,
  1367. struct mlxsw_sp_ipip_entry *ipip_entry,
  1368. struct net_device *ul_dev,
  1369. struct netlink_ext_ack *extack)
  1370. {
  1371. return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
  1372. true, true, false, extack);
  1373. }
  1374. static int
  1375. mlxsw_sp_netdevice_ipip_ul_up_event(struct mlxsw_sp *mlxsw_sp,
  1376. struct mlxsw_sp_ipip_entry *ipip_entry,
  1377. struct net_device *ul_dev)
  1378. {
  1379. return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
  1380. false, false, true, NULL);
  1381. }
  1382. static int
  1383. mlxsw_sp_netdevice_ipip_ul_down_event(struct mlxsw_sp *mlxsw_sp,
  1384. struct mlxsw_sp_ipip_entry *ipip_entry,
  1385. struct net_device *ul_dev)
  1386. {
  1387. /* A down underlay device causes encapsulated packets to not be
  1388. * forwarded, but decap still works. So refresh next hops without
  1389. * touching anything else.
  1390. */
  1391. return __mlxsw_sp_ipip_entry_update_tunnel(mlxsw_sp, ipip_entry,
  1392. false, false, true, NULL);
  1393. }
  1394. static int
  1395. mlxsw_sp_netdevice_ipip_ol_change_event(struct mlxsw_sp *mlxsw_sp,
  1396. struct net_device *ol_dev,
  1397. struct netlink_ext_ack *extack)
  1398. {
  1399. const struct mlxsw_sp_ipip_ops *ipip_ops;
  1400. struct mlxsw_sp_ipip_entry *ipip_entry;
  1401. int err;
  1402. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, ol_dev);
  1403. if (!ipip_entry)
  1404. /* A change might make a tunnel eligible for offloading, but
  1405. * that is currently not implemented. What falls to slow path
  1406. * stays there.
  1407. */
  1408. return 0;
  1409. /* A change might make a tunnel not eligible for offloading. */
  1410. if (!mlxsw_sp_netdevice_ipip_can_offload(mlxsw_sp, ol_dev,
  1411. ipip_entry->ipipt)) {
  1412. mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
  1413. return 0;
  1414. }
  1415. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
  1416. err = ipip_ops->ol_netdev_change(mlxsw_sp, ipip_entry, extack);
  1417. return err;
  1418. }
  1419. void mlxsw_sp_ipip_entry_demote_tunnel(struct mlxsw_sp *mlxsw_sp,
  1420. struct mlxsw_sp_ipip_entry *ipip_entry)
  1421. {
  1422. struct net_device *ol_dev = ipip_entry->ol_dev;
  1423. if (ol_dev->flags & IFF_UP)
  1424. mlxsw_sp_ipip_entry_ol_down_event(mlxsw_sp, ipip_entry);
  1425. mlxsw_sp_ipip_entry_destroy(mlxsw_sp, ipip_entry);
  1426. }
  1427. /* The configuration where several tunnels have the same local address in the
  1428. * same underlay table needs special treatment in the HW. That is currently not
  1429. * implemented in the driver. This function finds and demotes the first tunnel
  1430. * with a given source address, except the one passed in in the argument
  1431. * `except'.
  1432. */
  1433. bool
  1434. mlxsw_sp_ipip_demote_tunnel_by_saddr(struct mlxsw_sp *mlxsw_sp,
  1435. enum mlxsw_sp_l3proto ul_proto,
  1436. union mlxsw_sp_l3addr saddr,
  1437. u32 ul_tb_id,
  1438. const struct mlxsw_sp_ipip_entry *except)
  1439. {
  1440. struct mlxsw_sp_ipip_entry *ipip_entry, *tmp;
  1441. list_for_each_entry_safe(ipip_entry, tmp, &mlxsw_sp->router->ipip_list,
  1442. ipip_list_node) {
  1443. if (ipip_entry != except &&
  1444. mlxsw_sp_ipip_entry_saddr_matches(mlxsw_sp, ul_proto, saddr,
  1445. ul_tb_id, ipip_entry)) {
  1446. mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
  1447. return true;
  1448. }
  1449. }
  1450. return false;
  1451. }
  1452. static void mlxsw_sp_ipip_demote_tunnel_by_ul_netdev(struct mlxsw_sp *mlxsw_sp,
  1453. struct net_device *ul_dev)
  1454. {
  1455. struct mlxsw_sp_ipip_entry *ipip_entry, *tmp;
  1456. list_for_each_entry_safe(ipip_entry, tmp, &mlxsw_sp->router->ipip_list,
  1457. ipip_list_node) {
  1458. struct net_device *ipip_ul_dev =
  1459. __mlxsw_sp_ipip_netdev_ul_dev_get(ipip_entry->ol_dev);
  1460. if (ipip_ul_dev == ul_dev)
  1461. mlxsw_sp_ipip_entry_demote_tunnel(mlxsw_sp, ipip_entry);
  1462. }
  1463. }
  1464. int mlxsw_sp_netdevice_ipip_ol_event(struct mlxsw_sp *mlxsw_sp,
  1465. struct net_device *ol_dev,
  1466. unsigned long event,
  1467. struct netdev_notifier_info *info)
  1468. {
  1469. struct netdev_notifier_changeupper_info *chup;
  1470. struct netlink_ext_ack *extack;
  1471. switch (event) {
  1472. case NETDEV_REGISTER:
  1473. return mlxsw_sp_netdevice_ipip_ol_reg_event(mlxsw_sp, ol_dev);
  1474. case NETDEV_UNREGISTER:
  1475. mlxsw_sp_netdevice_ipip_ol_unreg_event(mlxsw_sp, ol_dev);
  1476. return 0;
  1477. case NETDEV_UP:
  1478. mlxsw_sp_netdevice_ipip_ol_up_event(mlxsw_sp, ol_dev);
  1479. return 0;
  1480. case NETDEV_DOWN:
  1481. mlxsw_sp_netdevice_ipip_ol_down_event(mlxsw_sp, ol_dev);
  1482. return 0;
  1483. case NETDEV_CHANGEUPPER:
  1484. chup = container_of(info, typeof(*chup), info);
  1485. extack = info->extack;
  1486. if (netif_is_l3_master(chup->upper_dev))
  1487. return mlxsw_sp_netdevice_ipip_ol_vrf_event(mlxsw_sp,
  1488. ol_dev,
  1489. extack);
  1490. return 0;
  1491. case NETDEV_CHANGE:
  1492. extack = info->extack;
  1493. return mlxsw_sp_netdevice_ipip_ol_change_event(mlxsw_sp,
  1494. ol_dev, extack);
  1495. case NETDEV_CHANGEMTU:
  1496. return mlxsw_sp_netdevice_ipip_ol_update_mtu(mlxsw_sp, ol_dev);
  1497. }
  1498. return 0;
  1499. }
  1500. static int
  1501. __mlxsw_sp_netdevice_ipip_ul_event(struct mlxsw_sp *mlxsw_sp,
  1502. struct mlxsw_sp_ipip_entry *ipip_entry,
  1503. struct net_device *ul_dev,
  1504. unsigned long event,
  1505. struct netdev_notifier_info *info)
  1506. {
  1507. struct netdev_notifier_changeupper_info *chup;
  1508. struct netlink_ext_ack *extack;
  1509. switch (event) {
  1510. case NETDEV_CHANGEUPPER:
  1511. chup = container_of(info, typeof(*chup), info);
  1512. extack = info->extack;
  1513. if (netif_is_l3_master(chup->upper_dev))
  1514. return mlxsw_sp_netdevice_ipip_ul_vrf_event(mlxsw_sp,
  1515. ipip_entry,
  1516. ul_dev,
  1517. extack);
  1518. break;
  1519. case NETDEV_UP:
  1520. return mlxsw_sp_netdevice_ipip_ul_up_event(mlxsw_sp, ipip_entry,
  1521. ul_dev);
  1522. case NETDEV_DOWN:
  1523. return mlxsw_sp_netdevice_ipip_ul_down_event(mlxsw_sp,
  1524. ipip_entry,
  1525. ul_dev);
  1526. }
  1527. return 0;
  1528. }
  1529. int
  1530. mlxsw_sp_netdevice_ipip_ul_event(struct mlxsw_sp *mlxsw_sp,
  1531. struct net_device *ul_dev,
  1532. unsigned long event,
  1533. struct netdev_notifier_info *info)
  1534. {
  1535. struct mlxsw_sp_ipip_entry *ipip_entry = NULL;
  1536. int err;
  1537. while ((ipip_entry = mlxsw_sp_ipip_entry_find_by_ul_dev(mlxsw_sp,
  1538. ul_dev,
  1539. ipip_entry))) {
  1540. err = __mlxsw_sp_netdevice_ipip_ul_event(mlxsw_sp, ipip_entry,
  1541. ul_dev, event, info);
  1542. if (err) {
  1543. mlxsw_sp_ipip_demote_tunnel_by_ul_netdev(mlxsw_sp,
  1544. ul_dev);
  1545. return err;
  1546. }
  1547. }
  1548. return 0;
  1549. }
  1550. struct mlxsw_sp_neigh_key {
  1551. struct neighbour *n;
  1552. };
  1553. struct mlxsw_sp_neigh_entry {
  1554. struct list_head rif_list_node;
  1555. struct rhash_head ht_node;
  1556. struct mlxsw_sp_neigh_key key;
  1557. u16 rif;
  1558. bool connected;
  1559. unsigned char ha[ETH_ALEN];
  1560. struct list_head nexthop_list; /* list of nexthops using
  1561. * this neigh entry
  1562. */
  1563. struct list_head nexthop_neighs_list_node;
  1564. unsigned int counter_index;
  1565. bool counter_valid;
  1566. };
  1567. static const struct rhashtable_params mlxsw_sp_neigh_ht_params = {
  1568. .key_offset = offsetof(struct mlxsw_sp_neigh_entry, key),
  1569. .head_offset = offsetof(struct mlxsw_sp_neigh_entry, ht_node),
  1570. .key_len = sizeof(struct mlxsw_sp_neigh_key),
  1571. };
  1572. struct mlxsw_sp_neigh_entry *
  1573. mlxsw_sp_rif_neigh_next(struct mlxsw_sp_rif *rif,
  1574. struct mlxsw_sp_neigh_entry *neigh_entry)
  1575. {
  1576. if (!neigh_entry) {
  1577. if (list_empty(&rif->neigh_list))
  1578. return NULL;
  1579. else
  1580. return list_first_entry(&rif->neigh_list,
  1581. typeof(*neigh_entry),
  1582. rif_list_node);
  1583. }
  1584. if (list_is_last(&neigh_entry->rif_list_node, &rif->neigh_list))
  1585. return NULL;
  1586. return list_next_entry(neigh_entry, rif_list_node);
  1587. }
  1588. int mlxsw_sp_neigh_entry_type(struct mlxsw_sp_neigh_entry *neigh_entry)
  1589. {
  1590. return neigh_entry->key.n->tbl->family;
  1591. }
  1592. unsigned char *
  1593. mlxsw_sp_neigh_entry_ha(struct mlxsw_sp_neigh_entry *neigh_entry)
  1594. {
  1595. return neigh_entry->ha;
  1596. }
  1597. u32 mlxsw_sp_neigh4_entry_dip(struct mlxsw_sp_neigh_entry *neigh_entry)
  1598. {
  1599. struct neighbour *n;
  1600. n = neigh_entry->key.n;
  1601. return ntohl(*((__be32 *) n->primary_key));
  1602. }
  1603. struct in6_addr *
  1604. mlxsw_sp_neigh6_entry_dip(struct mlxsw_sp_neigh_entry *neigh_entry)
  1605. {
  1606. struct neighbour *n;
  1607. n = neigh_entry->key.n;
  1608. return (struct in6_addr *) &n->primary_key;
  1609. }
  1610. int mlxsw_sp_neigh_counter_get(struct mlxsw_sp *mlxsw_sp,
  1611. struct mlxsw_sp_neigh_entry *neigh_entry,
  1612. u64 *p_counter)
  1613. {
  1614. if (!neigh_entry->counter_valid)
  1615. return -EINVAL;
  1616. return mlxsw_sp_flow_counter_get(mlxsw_sp, neigh_entry->counter_index,
  1617. p_counter, NULL);
  1618. }
  1619. static struct mlxsw_sp_neigh_entry *
  1620. mlxsw_sp_neigh_entry_alloc(struct mlxsw_sp *mlxsw_sp, struct neighbour *n,
  1621. u16 rif)
  1622. {
  1623. struct mlxsw_sp_neigh_entry *neigh_entry;
  1624. neigh_entry = kzalloc(sizeof(*neigh_entry), GFP_KERNEL);
  1625. if (!neigh_entry)
  1626. return NULL;
  1627. neigh_entry->key.n = n;
  1628. neigh_entry->rif = rif;
  1629. INIT_LIST_HEAD(&neigh_entry->nexthop_list);
  1630. return neigh_entry;
  1631. }
  1632. static void mlxsw_sp_neigh_entry_free(struct mlxsw_sp_neigh_entry *neigh_entry)
  1633. {
  1634. kfree(neigh_entry);
  1635. }
  1636. static int
  1637. mlxsw_sp_neigh_entry_insert(struct mlxsw_sp *mlxsw_sp,
  1638. struct mlxsw_sp_neigh_entry *neigh_entry)
  1639. {
  1640. return rhashtable_insert_fast(&mlxsw_sp->router->neigh_ht,
  1641. &neigh_entry->ht_node,
  1642. mlxsw_sp_neigh_ht_params);
  1643. }
  1644. static void
  1645. mlxsw_sp_neigh_entry_remove(struct mlxsw_sp *mlxsw_sp,
  1646. struct mlxsw_sp_neigh_entry *neigh_entry)
  1647. {
  1648. rhashtable_remove_fast(&mlxsw_sp->router->neigh_ht,
  1649. &neigh_entry->ht_node,
  1650. mlxsw_sp_neigh_ht_params);
  1651. }
  1652. static bool
  1653. mlxsw_sp_neigh_counter_should_alloc(struct mlxsw_sp *mlxsw_sp,
  1654. struct mlxsw_sp_neigh_entry *neigh_entry)
  1655. {
  1656. struct devlink *devlink;
  1657. const char *table_name;
  1658. switch (mlxsw_sp_neigh_entry_type(neigh_entry)) {
  1659. case AF_INET:
  1660. table_name = MLXSW_SP_DPIPE_TABLE_NAME_HOST4;
  1661. break;
  1662. case AF_INET6:
  1663. table_name = MLXSW_SP_DPIPE_TABLE_NAME_HOST6;
  1664. break;
  1665. default:
  1666. WARN_ON(1);
  1667. return false;
  1668. }
  1669. devlink = priv_to_devlink(mlxsw_sp->core);
  1670. return devlink_dpipe_table_counter_enabled(devlink, table_name);
  1671. }
  1672. static void
  1673. mlxsw_sp_neigh_counter_alloc(struct mlxsw_sp *mlxsw_sp,
  1674. struct mlxsw_sp_neigh_entry *neigh_entry)
  1675. {
  1676. if (!mlxsw_sp_neigh_counter_should_alloc(mlxsw_sp, neigh_entry))
  1677. return;
  1678. if (mlxsw_sp_flow_counter_alloc(mlxsw_sp, &neigh_entry->counter_index))
  1679. return;
  1680. neigh_entry->counter_valid = true;
  1681. }
  1682. static void
  1683. mlxsw_sp_neigh_counter_free(struct mlxsw_sp *mlxsw_sp,
  1684. struct mlxsw_sp_neigh_entry *neigh_entry)
  1685. {
  1686. if (!neigh_entry->counter_valid)
  1687. return;
  1688. mlxsw_sp_flow_counter_free(mlxsw_sp,
  1689. neigh_entry->counter_index);
  1690. neigh_entry->counter_valid = false;
  1691. }
  1692. static struct mlxsw_sp_neigh_entry *
  1693. mlxsw_sp_neigh_entry_create(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
  1694. {
  1695. struct mlxsw_sp_neigh_entry *neigh_entry;
  1696. struct mlxsw_sp_rif *rif;
  1697. int err;
  1698. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, n->dev);
  1699. if (!rif)
  1700. return ERR_PTR(-EINVAL);
  1701. neigh_entry = mlxsw_sp_neigh_entry_alloc(mlxsw_sp, n, rif->rif_index);
  1702. if (!neigh_entry)
  1703. return ERR_PTR(-ENOMEM);
  1704. err = mlxsw_sp_neigh_entry_insert(mlxsw_sp, neigh_entry);
  1705. if (err)
  1706. goto err_neigh_entry_insert;
  1707. mlxsw_sp_neigh_counter_alloc(mlxsw_sp, neigh_entry);
  1708. list_add(&neigh_entry->rif_list_node, &rif->neigh_list);
  1709. return neigh_entry;
  1710. err_neigh_entry_insert:
  1711. mlxsw_sp_neigh_entry_free(neigh_entry);
  1712. return ERR_PTR(err);
  1713. }
  1714. static void
  1715. mlxsw_sp_neigh_entry_destroy(struct mlxsw_sp *mlxsw_sp,
  1716. struct mlxsw_sp_neigh_entry *neigh_entry)
  1717. {
  1718. list_del(&neigh_entry->rif_list_node);
  1719. mlxsw_sp_neigh_counter_free(mlxsw_sp, neigh_entry);
  1720. mlxsw_sp_neigh_entry_remove(mlxsw_sp, neigh_entry);
  1721. mlxsw_sp_neigh_entry_free(neigh_entry);
  1722. }
  1723. static struct mlxsw_sp_neigh_entry *
  1724. mlxsw_sp_neigh_entry_lookup(struct mlxsw_sp *mlxsw_sp, struct neighbour *n)
  1725. {
  1726. struct mlxsw_sp_neigh_key key;
  1727. key.n = n;
  1728. return rhashtable_lookup_fast(&mlxsw_sp->router->neigh_ht,
  1729. &key, mlxsw_sp_neigh_ht_params);
  1730. }
  1731. static void
  1732. mlxsw_sp_router_neighs_update_interval_init(struct mlxsw_sp *mlxsw_sp)
  1733. {
  1734. unsigned long interval;
  1735. #if IS_ENABLED(CONFIG_IPV6)
  1736. interval = min_t(unsigned long,
  1737. NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME),
  1738. NEIGH_VAR(&nd_tbl.parms, DELAY_PROBE_TIME));
  1739. #else
  1740. interval = NEIGH_VAR(&arp_tbl.parms, DELAY_PROBE_TIME);
  1741. #endif
  1742. mlxsw_sp->router->neighs_update.interval = jiffies_to_msecs(interval);
  1743. }
  1744. static void mlxsw_sp_router_neigh_ent_ipv4_process(struct mlxsw_sp *mlxsw_sp,
  1745. char *rauhtd_pl,
  1746. int ent_index)
  1747. {
  1748. struct net_device *dev;
  1749. struct neighbour *n;
  1750. __be32 dipn;
  1751. u32 dip;
  1752. u16 rif;
  1753. mlxsw_reg_rauhtd_ent_ipv4_unpack(rauhtd_pl, ent_index, &rif, &dip);
  1754. if (!mlxsw_sp->router->rifs[rif]) {
  1755. dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
  1756. return;
  1757. }
  1758. dipn = htonl(dip);
  1759. dev = mlxsw_sp->router->rifs[rif]->dev;
  1760. n = neigh_lookup(&arp_tbl, &dipn, dev);
  1761. if (!n)
  1762. return;
  1763. netdev_dbg(dev, "Updating neighbour with IP=%pI4h\n", &dip);
  1764. neigh_event_send(n, NULL);
  1765. neigh_release(n);
  1766. }
  1767. #if IS_ENABLED(CONFIG_IPV6)
  1768. static void mlxsw_sp_router_neigh_ent_ipv6_process(struct mlxsw_sp *mlxsw_sp,
  1769. char *rauhtd_pl,
  1770. int rec_index)
  1771. {
  1772. struct net_device *dev;
  1773. struct neighbour *n;
  1774. struct in6_addr dip;
  1775. u16 rif;
  1776. mlxsw_reg_rauhtd_ent_ipv6_unpack(rauhtd_pl, rec_index, &rif,
  1777. (char *) &dip);
  1778. if (!mlxsw_sp->router->rifs[rif]) {
  1779. dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Incorrect RIF in neighbour entry\n");
  1780. return;
  1781. }
  1782. dev = mlxsw_sp->router->rifs[rif]->dev;
  1783. n = neigh_lookup(&nd_tbl, &dip, dev);
  1784. if (!n)
  1785. return;
  1786. netdev_dbg(dev, "Updating neighbour with IP=%pI6c\n", &dip);
  1787. neigh_event_send(n, NULL);
  1788. neigh_release(n);
  1789. }
  1790. #else
  1791. static void mlxsw_sp_router_neigh_ent_ipv6_process(struct mlxsw_sp *mlxsw_sp,
  1792. char *rauhtd_pl,
  1793. int rec_index)
  1794. {
  1795. }
  1796. #endif
  1797. static void mlxsw_sp_router_neigh_rec_ipv4_process(struct mlxsw_sp *mlxsw_sp,
  1798. char *rauhtd_pl,
  1799. int rec_index)
  1800. {
  1801. u8 num_entries;
  1802. int i;
  1803. num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
  1804. rec_index);
  1805. /* Hardware starts counting at 0, so add 1. */
  1806. num_entries++;
  1807. /* Each record consists of several neighbour entries. */
  1808. for (i = 0; i < num_entries; i++) {
  1809. int ent_index;
  1810. ent_index = rec_index * MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC + i;
  1811. mlxsw_sp_router_neigh_ent_ipv4_process(mlxsw_sp, rauhtd_pl,
  1812. ent_index);
  1813. }
  1814. }
  1815. static void mlxsw_sp_router_neigh_rec_ipv6_process(struct mlxsw_sp *mlxsw_sp,
  1816. char *rauhtd_pl,
  1817. int rec_index)
  1818. {
  1819. /* One record contains one entry. */
  1820. mlxsw_sp_router_neigh_ent_ipv6_process(mlxsw_sp, rauhtd_pl,
  1821. rec_index);
  1822. }
  1823. static void mlxsw_sp_router_neigh_rec_process(struct mlxsw_sp *mlxsw_sp,
  1824. char *rauhtd_pl, int rec_index)
  1825. {
  1826. switch (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, rec_index)) {
  1827. case MLXSW_REG_RAUHTD_TYPE_IPV4:
  1828. mlxsw_sp_router_neigh_rec_ipv4_process(mlxsw_sp, rauhtd_pl,
  1829. rec_index);
  1830. break;
  1831. case MLXSW_REG_RAUHTD_TYPE_IPV6:
  1832. mlxsw_sp_router_neigh_rec_ipv6_process(mlxsw_sp, rauhtd_pl,
  1833. rec_index);
  1834. break;
  1835. }
  1836. }
  1837. static bool mlxsw_sp_router_rauhtd_is_full(char *rauhtd_pl)
  1838. {
  1839. u8 num_rec, last_rec_index, num_entries;
  1840. num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
  1841. last_rec_index = num_rec - 1;
  1842. if (num_rec < MLXSW_REG_RAUHTD_REC_MAX_NUM)
  1843. return false;
  1844. if (mlxsw_reg_rauhtd_rec_type_get(rauhtd_pl, last_rec_index) ==
  1845. MLXSW_REG_RAUHTD_TYPE_IPV6)
  1846. return true;
  1847. num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
  1848. last_rec_index);
  1849. if (++num_entries == MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC)
  1850. return true;
  1851. return false;
  1852. }
  1853. static int
  1854. __mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp,
  1855. char *rauhtd_pl,
  1856. enum mlxsw_reg_rauhtd_type type)
  1857. {
  1858. int i, num_rec;
  1859. int err;
  1860. /* Make sure the neighbour's netdev isn't removed in the
  1861. * process.
  1862. */
  1863. rtnl_lock();
  1864. do {
  1865. mlxsw_reg_rauhtd_pack(rauhtd_pl, type);
  1866. err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(rauhtd),
  1867. rauhtd_pl);
  1868. if (err) {
  1869. dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to dump neighbour table\n");
  1870. break;
  1871. }
  1872. num_rec = mlxsw_reg_rauhtd_num_rec_get(rauhtd_pl);
  1873. for (i = 0; i < num_rec; i++)
  1874. mlxsw_sp_router_neigh_rec_process(mlxsw_sp, rauhtd_pl,
  1875. i);
  1876. } while (mlxsw_sp_router_rauhtd_is_full(rauhtd_pl));
  1877. rtnl_unlock();
  1878. return err;
  1879. }
  1880. static int mlxsw_sp_router_neighs_update_rauhtd(struct mlxsw_sp *mlxsw_sp)
  1881. {
  1882. enum mlxsw_reg_rauhtd_type type;
  1883. char *rauhtd_pl;
  1884. int err;
  1885. rauhtd_pl = kmalloc(MLXSW_REG_RAUHTD_LEN, GFP_KERNEL);
  1886. if (!rauhtd_pl)
  1887. return -ENOMEM;
  1888. type = MLXSW_REG_RAUHTD_TYPE_IPV4;
  1889. err = __mlxsw_sp_router_neighs_update_rauhtd(mlxsw_sp, rauhtd_pl, type);
  1890. if (err)
  1891. goto out;
  1892. type = MLXSW_REG_RAUHTD_TYPE_IPV6;
  1893. err = __mlxsw_sp_router_neighs_update_rauhtd(mlxsw_sp, rauhtd_pl, type);
  1894. out:
  1895. kfree(rauhtd_pl);
  1896. return err;
  1897. }
  1898. static void mlxsw_sp_router_neighs_update_nh(struct mlxsw_sp *mlxsw_sp)
  1899. {
  1900. struct mlxsw_sp_neigh_entry *neigh_entry;
  1901. /* Take RTNL mutex here to prevent lists from changes */
  1902. rtnl_lock();
  1903. list_for_each_entry(neigh_entry, &mlxsw_sp->router->nexthop_neighs_list,
  1904. nexthop_neighs_list_node)
  1905. /* If this neigh have nexthops, make the kernel think this neigh
  1906. * is active regardless of the traffic.
  1907. */
  1908. neigh_event_send(neigh_entry->key.n, NULL);
  1909. rtnl_unlock();
  1910. }
  1911. static void
  1912. mlxsw_sp_router_neighs_update_work_schedule(struct mlxsw_sp *mlxsw_sp)
  1913. {
  1914. unsigned long interval = mlxsw_sp->router->neighs_update.interval;
  1915. mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw,
  1916. msecs_to_jiffies(interval));
  1917. }
  1918. static void mlxsw_sp_router_neighs_update_work(struct work_struct *work)
  1919. {
  1920. struct mlxsw_sp_router *router;
  1921. int err;
  1922. router = container_of(work, struct mlxsw_sp_router,
  1923. neighs_update.dw.work);
  1924. err = mlxsw_sp_router_neighs_update_rauhtd(router->mlxsw_sp);
  1925. if (err)
  1926. dev_err(router->mlxsw_sp->bus_info->dev, "Could not update kernel for neigh activity");
  1927. mlxsw_sp_router_neighs_update_nh(router->mlxsw_sp);
  1928. mlxsw_sp_router_neighs_update_work_schedule(router->mlxsw_sp);
  1929. }
  1930. static void mlxsw_sp_router_probe_unresolved_nexthops(struct work_struct *work)
  1931. {
  1932. struct mlxsw_sp_neigh_entry *neigh_entry;
  1933. struct mlxsw_sp_router *router;
  1934. router = container_of(work, struct mlxsw_sp_router,
  1935. nexthop_probe_dw.work);
  1936. /* Iterate over nexthop neighbours, find those who are unresolved and
  1937. * send arp on them. This solves the chicken-egg problem when
  1938. * the nexthop wouldn't get offloaded until the neighbor is resolved
  1939. * but it wouldn't get resolved ever in case traffic is flowing in HW
  1940. * using different nexthop.
  1941. *
  1942. * Take RTNL mutex here to prevent lists from changes.
  1943. */
  1944. rtnl_lock();
  1945. list_for_each_entry(neigh_entry, &router->nexthop_neighs_list,
  1946. nexthop_neighs_list_node)
  1947. if (!neigh_entry->connected)
  1948. neigh_event_send(neigh_entry->key.n, NULL);
  1949. rtnl_unlock();
  1950. mlxsw_core_schedule_dw(&router->nexthop_probe_dw,
  1951. MLXSW_SP_UNRESOLVED_NH_PROBE_INTERVAL);
  1952. }
  1953. static void
  1954. mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
  1955. struct mlxsw_sp_neigh_entry *neigh_entry,
  1956. bool removing);
  1957. static enum mlxsw_reg_rauht_op mlxsw_sp_rauht_op(bool adding)
  1958. {
  1959. return adding ? MLXSW_REG_RAUHT_OP_WRITE_ADD :
  1960. MLXSW_REG_RAUHT_OP_WRITE_DELETE;
  1961. }
  1962. static void
  1963. mlxsw_sp_router_neigh_entry_op4(struct mlxsw_sp *mlxsw_sp,
  1964. struct mlxsw_sp_neigh_entry *neigh_entry,
  1965. enum mlxsw_reg_rauht_op op)
  1966. {
  1967. struct neighbour *n = neigh_entry->key.n;
  1968. u32 dip = ntohl(*((__be32 *) n->primary_key));
  1969. char rauht_pl[MLXSW_REG_RAUHT_LEN];
  1970. mlxsw_reg_rauht_pack4(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
  1971. dip);
  1972. if (neigh_entry->counter_valid)
  1973. mlxsw_reg_rauht_pack_counter(rauht_pl,
  1974. neigh_entry->counter_index);
  1975. mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
  1976. }
  1977. static void
  1978. mlxsw_sp_router_neigh_entry_op6(struct mlxsw_sp *mlxsw_sp,
  1979. struct mlxsw_sp_neigh_entry *neigh_entry,
  1980. enum mlxsw_reg_rauht_op op)
  1981. {
  1982. struct neighbour *n = neigh_entry->key.n;
  1983. char rauht_pl[MLXSW_REG_RAUHT_LEN];
  1984. const char *dip = n->primary_key;
  1985. mlxsw_reg_rauht_pack6(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
  1986. dip);
  1987. if (neigh_entry->counter_valid)
  1988. mlxsw_reg_rauht_pack_counter(rauht_pl,
  1989. neigh_entry->counter_index);
  1990. mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rauht), rauht_pl);
  1991. }
  1992. bool mlxsw_sp_neigh_ipv6_ignore(struct mlxsw_sp_neigh_entry *neigh_entry)
  1993. {
  1994. struct neighbour *n = neigh_entry->key.n;
  1995. /* Packets with a link-local destination address are trapped
  1996. * after LPM lookup and never reach the neighbour table, so
  1997. * there is no need to program such neighbours to the device.
  1998. */
  1999. if (ipv6_addr_type((struct in6_addr *) &n->primary_key) &
  2000. IPV6_ADDR_LINKLOCAL)
  2001. return true;
  2002. return false;
  2003. }
  2004. static void
  2005. mlxsw_sp_neigh_entry_update(struct mlxsw_sp *mlxsw_sp,
  2006. struct mlxsw_sp_neigh_entry *neigh_entry,
  2007. bool adding)
  2008. {
  2009. if (!adding && !neigh_entry->connected)
  2010. return;
  2011. neigh_entry->connected = adding;
  2012. if (neigh_entry->key.n->tbl->family == AF_INET) {
  2013. mlxsw_sp_router_neigh_entry_op4(mlxsw_sp, neigh_entry,
  2014. mlxsw_sp_rauht_op(adding));
  2015. } else if (neigh_entry->key.n->tbl->family == AF_INET6) {
  2016. if (mlxsw_sp_neigh_ipv6_ignore(neigh_entry))
  2017. return;
  2018. mlxsw_sp_router_neigh_entry_op6(mlxsw_sp, neigh_entry,
  2019. mlxsw_sp_rauht_op(adding));
  2020. } else {
  2021. WARN_ON_ONCE(1);
  2022. }
  2023. }
  2024. void
  2025. mlxsw_sp_neigh_entry_counter_update(struct mlxsw_sp *mlxsw_sp,
  2026. struct mlxsw_sp_neigh_entry *neigh_entry,
  2027. bool adding)
  2028. {
  2029. if (adding)
  2030. mlxsw_sp_neigh_counter_alloc(mlxsw_sp, neigh_entry);
  2031. else
  2032. mlxsw_sp_neigh_counter_free(mlxsw_sp, neigh_entry);
  2033. mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, true);
  2034. }
  2035. struct mlxsw_sp_netevent_work {
  2036. struct work_struct work;
  2037. struct mlxsw_sp *mlxsw_sp;
  2038. struct neighbour *n;
  2039. };
  2040. static void mlxsw_sp_router_neigh_event_work(struct work_struct *work)
  2041. {
  2042. struct mlxsw_sp_netevent_work *net_work =
  2043. container_of(work, struct mlxsw_sp_netevent_work, work);
  2044. struct mlxsw_sp *mlxsw_sp = net_work->mlxsw_sp;
  2045. struct mlxsw_sp_neigh_entry *neigh_entry;
  2046. struct neighbour *n = net_work->n;
  2047. unsigned char ha[ETH_ALEN];
  2048. bool entry_connected;
  2049. u8 nud_state, dead;
  2050. /* If these parameters are changed after we release the lock,
  2051. * then we are guaranteed to receive another event letting us
  2052. * know about it.
  2053. */
  2054. read_lock_bh(&n->lock);
  2055. memcpy(ha, n->ha, ETH_ALEN);
  2056. nud_state = n->nud_state;
  2057. dead = n->dead;
  2058. read_unlock_bh(&n->lock);
  2059. rtnl_lock();
  2060. mlxsw_sp_span_respin(mlxsw_sp);
  2061. entry_connected = nud_state & NUD_VALID && !dead;
  2062. neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
  2063. if (!entry_connected && !neigh_entry)
  2064. goto out;
  2065. if (!neigh_entry) {
  2066. neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
  2067. if (IS_ERR(neigh_entry))
  2068. goto out;
  2069. }
  2070. memcpy(neigh_entry->ha, ha, ETH_ALEN);
  2071. mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, entry_connected);
  2072. mlxsw_sp_nexthop_neigh_update(mlxsw_sp, neigh_entry, !entry_connected);
  2073. if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
  2074. mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
  2075. out:
  2076. rtnl_unlock();
  2077. neigh_release(n);
  2078. kfree(net_work);
  2079. }
  2080. static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp);
  2081. static void mlxsw_sp_router_mp_hash_event_work(struct work_struct *work)
  2082. {
  2083. struct mlxsw_sp_netevent_work *net_work =
  2084. container_of(work, struct mlxsw_sp_netevent_work, work);
  2085. struct mlxsw_sp *mlxsw_sp = net_work->mlxsw_sp;
  2086. mlxsw_sp_mp_hash_init(mlxsw_sp);
  2087. kfree(net_work);
  2088. }
  2089. static int mlxsw_sp_router_netevent_event(struct notifier_block *nb,
  2090. unsigned long event, void *ptr)
  2091. {
  2092. struct mlxsw_sp_netevent_work *net_work;
  2093. struct mlxsw_sp_port *mlxsw_sp_port;
  2094. struct mlxsw_sp_router *router;
  2095. struct mlxsw_sp *mlxsw_sp;
  2096. unsigned long interval;
  2097. struct neigh_parms *p;
  2098. struct neighbour *n;
  2099. struct net *net;
  2100. switch (event) {
  2101. case NETEVENT_DELAY_PROBE_TIME_UPDATE:
  2102. p = ptr;
  2103. /* We don't care about changes in the default table. */
  2104. if (!p->dev || (p->tbl->family != AF_INET &&
  2105. p->tbl->family != AF_INET6))
  2106. return NOTIFY_DONE;
  2107. /* We are in atomic context and can't take RTNL mutex,
  2108. * so use RCU variant to walk the device chain.
  2109. */
  2110. mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(p->dev);
  2111. if (!mlxsw_sp_port)
  2112. return NOTIFY_DONE;
  2113. mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
  2114. interval = jiffies_to_msecs(NEIGH_VAR(p, DELAY_PROBE_TIME));
  2115. mlxsw_sp->router->neighs_update.interval = interval;
  2116. mlxsw_sp_port_dev_put(mlxsw_sp_port);
  2117. break;
  2118. case NETEVENT_NEIGH_UPDATE:
  2119. n = ptr;
  2120. if (n->tbl->family != AF_INET && n->tbl->family != AF_INET6)
  2121. return NOTIFY_DONE;
  2122. mlxsw_sp_port = mlxsw_sp_port_lower_dev_hold(n->dev);
  2123. if (!mlxsw_sp_port)
  2124. return NOTIFY_DONE;
  2125. net_work = kzalloc(sizeof(*net_work), GFP_ATOMIC);
  2126. if (!net_work) {
  2127. mlxsw_sp_port_dev_put(mlxsw_sp_port);
  2128. return NOTIFY_BAD;
  2129. }
  2130. INIT_WORK(&net_work->work, mlxsw_sp_router_neigh_event_work);
  2131. net_work->mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
  2132. net_work->n = n;
  2133. /* Take a reference to ensure the neighbour won't be
  2134. * destructed until we drop the reference in delayed
  2135. * work.
  2136. */
  2137. neigh_clone(n);
  2138. mlxsw_core_schedule_work(&net_work->work);
  2139. mlxsw_sp_port_dev_put(mlxsw_sp_port);
  2140. break;
  2141. case NETEVENT_IPV4_MPATH_HASH_UPDATE:
  2142. case NETEVENT_IPV6_MPATH_HASH_UPDATE:
  2143. net = ptr;
  2144. if (!net_eq(net, &init_net))
  2145. return NOTIFY_DONE;
  2146. net_work = kzalloc(sizeof(*net_work), GFP_ATOMIC);
  2147. if (!net_work)
  2148. return NOTIFY_BAD;
  2149. router = container_of(nb, struct mlxsw_sp_router, netevent_nb);
  2150. INIT_WORK(&net_work->work, mlxsw_sp_router_mp_hash_event_work);
  2151. net_work->mlxsw_sp = router->mlxsw_sp;
  2152. mlxsw_core_schedule_work(&net_work->work);
  2153. break;
  2154. }
  2155. return NOTIFY_DONE;
  2156. }
  2157. static int mlxsw_sp_neigh_init(struct mlxsw_sp *mlxsw_sp)
  2158. {
  2159. int err;
  2160. err = rhashtable_init(&mlxsw_sp->router->neigh_ht,
  2161. &mlxsw_sp_neigh_ht_params);
  2162. if (err)
  2163. return err;
  2164. /* Initialize the polling interval according to the default
  2165. * table.
  2166. */
  2167. mlxsw_sp_router_neighs_update_interval_init(mlxsw_sp);
  2168. /* Create the delayed works for the activity_update */
  2169. INIT_DELAYED_WORK(&mlxsw_sp->router->neighs_update.dw,
  2170. mlxsw_sp_router_neighs_update_work);
  2171. INIT_DELAYED_WORK(&mlxsw_sp->router->nexthop_probe_dw,
  2172. mlxsw_sp_router_probe_unresolved_nexthops);
  2173. mlxsw_core_schedule_dw(&mlxsw_sp->router->neighs_update.dw, 0);
  2174. mlxsw_core_schedule_dw(&mlxsw_sp->router->nexthop_probe_dw, 0);
  2175. return 0;
  2176. }
  2177. static void mlxsw_sp_neigh_fini(struct mlxsw_sp *mlxsw_sp)
  2178. {
  2179. cancel_delayed_work_sync(&mlxsw_sp->router->neighs_update.dw);
  2180. cancel_delayed_work_sync(&mlxsw_sp->router->nexthop_probe_dw);
  2181. rhashtable_destroy(&mlxsw_sp->router->neigh_ht);
  2182. }
  2183. static void mlxsw_sp_neigh_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
  2184. struct mlxsw_sp_rif *rif)
  2185. {
  2186. struct mlxsw_sp_neigh_entry *neigh_entry, *tmp;
  2187. list_for_each_entry_safe(neigh_entry, tmp, &rif->neigh_list,
  2188. rif_list_node) {
  2189. mlxsw_sp_neigh_entry_update(mlxsw_sp, neigh_entry, false);
  2190. mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
  2191. }
  2192. }
  2193. enum mlxsw_sp_nexthop_type {
  2194. MLXSW_SP_NEXTHOP_TYPE_ETH,
  2195. MLXSW_SP_NEXTHOP_TYPE_IPIP,
  2196. };
  2197. struct mlxsw_sp_nexthop_key {
  2198. struct fib_nh *fib_nh;
  2199. };
  2200. struct mlxsw_sp_nexthop {
  2201. struct list_head neigh_list_node; /* member of neigh entry list */
  2202. struct list_head rif_list_node;
  2203. struct list_head router_list_node;
  2204. struct mlxsw_sp_nexthop_group *nh_grp; /* pointer back to the group
  2205. * this belongs to
  2206. */
  2207. struct rhash_head ht_node;
  2208. struct mlxsw_sp_nexthop_key key;
  2209. unsigned char gw_addr[sizeof(struct in6_addr)];
  2210. int ifindex;
  2211. int nh_weight;
  2212. int norm_nh_weight;
  2213. int num_adj_entries;
  2214. struct mlxsw_sp_rif *rif;
  2215. u8 should_offload:1, /* set indicates this neigh is connected and
  2216. * should be put to KVD linear area of this group.
  2217. */
  2218. offloaded:1, /* set in case the neigh is actually put into
  2219. * KVD linear area of this group.
  2220. */
  2221. update:1; /* set indicates that MAC of this neigh should be
  2222. * updated in HW
  2223. */
  2224. enum mlxsw_sp_nexthop_type type;
  2225. union {
  2226. struct mlxsw_sp_neigh_entry *neigh_entry;
  2227. struct mlxsw_sp_ipip_entry *ipip_entry;
  2228. };
  2229. unsigned int counter_index;
  2230. bool counter_valid;
  2231. };
  2232. struct mlxsw_sp_nexthop_group {
  2233. void *priv;
  2234. struct rhash_head ht_node;
  2235. struct list_head fib_list; /* list of fib entries that use this group */
  2236. struct neigh_table *neigh_tbl;
  2237. u8 adj_index_valid:1,
  2238. gateway:1; /* routes using the group use a gateway */
  2239. u32 adj_index;
  2240. u16 ecmp_size;
  2241. u16 count;
  2242. int sum_norm_weight;
  2243. struct mlxsw_sp_nexthop nexthops[0];
  2244. #define nh_rif nexthops[0].rif
  2245. };
  2246. void mlxsw_sp_nexthop_counter_alloc(struct mlxsw_sp *mlxsw_sp,
  2247. struct mlxsw_sp_nexthop *nh)
  2248. {
  2249. struct devlink *devlink;
  2250. devlink = priv_to_devlink(mlxsw_sp->core);
  2251. if (!devlink_dpipe_table_counter_enabled(devlink,
  2252. MLXSW_SP_DPIPE_TABLE_NAME_ADJ))
  2253. return;
  2254. if (mlxsw_sp_flow_counter_alloc(mlxsw_sp, &nh->counter_index))
  2255. return;
  2256. nh->counter_valid = true;
  2257. }
  2258. void mlxsw_sp_nexthop_counter_free(struct mlxsw_sp *mlxsw_sp,
  2259. struct mlxsw_sp_nexthop *nh)
  2260. {
  2261. if (!nh->counter_valid)
  2262. return;
  2263. mlxsw_sp_flow_counter_free(mlxsw_sp, nh->counter_index);
  2264. nh->counter_valid = false;
  2265. }
  2266. int mlxsw_sp_nexthop_counter_get(struct mlxsw_sp *mlxsw_sp,
  2267. struct mlxsw_sp_nexthop *nh, u64 *p_counter)
  2268. {
  2269. if (!nh->counter_valid)
  2270. return -EINVAL;
  2271. return mlxsw_sp_flow_counter_get(mlxsw_sp, nh->counter_index,
  2272. p_counter, NULL);
  2273. }
  2274. struct mlxsw_sp_nexthop *mlxsw_sp_nexthop_next(struct mlxsw_sp_router *router,
  2275. struct mlxsw_sp_nexthop *nh)
  2276. {
  2277. if (!nh) {
  2278. if (list_empty(&router->nexthop_list))
  2279. return NULL;
  2280. else
  2281. return list_first_entry(&router->nexthop_list,
  2282. typeof(*nh), router_list_node);
  2283. }
  2284. if (list_is_last(&nh->router_list_node, &router->nexthop_list))
  2285. return NULL;
  2286. return list_next_entry(nh, router_list_node);
  2287. }
  2288. bool mlxsw_sp_nexthop_offload(struct mlxsw_sp_nexthop *nh)
  2289. {
  2290. return nh->offloaded;
  2291. }
  2292. unsigned char *mlxsw_sp_nexthop_ha(struct mlxsw_sp_nexthop *nh)
  2293. {
  2294. if (!nh->offloaded)
  2295. return NULL;
  2296. return nh->neigh_entry->ha;
  2297. }
  2298. int mlxsw_sp_nexthop_indexes(struct mlxsw_sp_nexthop *nh, u32 *p_adj_index,
  2299. u32 *p_adj_size, u32 *p_adj_hash_index)
  2300. {
  2301. struct mlxsw_sp_nexthop_group *nh_grp = nh->nh_grp;
  2302. u32 adj_hash_index = 0;
  2303. int i;
  2304. if (!nh->offloaded || !nh_grp->adj_index_valid)
  2305. return -EINVAL;
  2306. *p_adj_index = nh_grp->adj_index;
  2307. *p_adj_size = nh_grp->ecmp_size;
  2308. for (i = 0; i < nh_grp->count; i++) {
  2309. struct mlxsw_sp_nexthop *nh_iter = &nh_grp->nexthops[i];
  2310. if (nh_iter == nh)
  2311. break;
  2312. if (nh_iter->offloaded)
  2313. adj_hash_index += nh_iter->num_adj_entries;
  2314. }
  2315. *p_adj_hash_index = adj_hash_index;
  2316. return 0;
  2317. }
  2318. struct mlxsw_sp_rif *mlxsw_sp_nexthop_rif(struct mlxsw_sp_nexthop *nh)
  2319. {
  2320. return nh->rif;
  2321. }
  2322. bool mlxsw_sp_nexthop_group_has_ipip(struct mlxsw_sp_nexthop *nh)
  2323. {
  2324. struct mlxsw_sp_nexthop_group *nh_grp = nh->nh_grp;
  2325. int i;
  2326. for (i = 0; i < nh_grp->count; i++) {
  2327. struct mlxsw_sp_nexthop *nh_iter = &nh_grp->nexthops[i];
  2328. if (nh_iter->type == MLXSW_SP_NEXTHOP_TYPE_IPIP)
  2329. return true;
  2330. }
  2331. return false;
  2332. }
  2333. static struct fib_info *
  2334. mlxsw_sp_nexthop4_group_fi(const struct mlxsw_sp_nexthop_group *nh_grp)
  2335. {
  2336. return nh_grp->priv;
  2337. }
  2338. struct mlxsw_sp_nexthop_group_cmp_arg {
  2339. enum mlxsw_sp_l3proto proto;
  2340. union {
  2341. struct fib_info *fi;
  2342. struct mlxsw_sp_fib6_entry *fib6_entry;
  2343. };
  2344. };
  2345. static bool
  2346. mlxsw_sp_nexthop6_group_has_nexthop(const struct mlxsw_sp_nexthop_group *nh_grp,
  2347. const struct in6_addr *gw, int ifindex,
  2348. int weight)
  2349. {
  2350. int i;
  2351. for (i = 0; i < nh_grp->count; i++) {
  2352. const struct mlxsw_sp_nexthop *nh;
  2353. nh = &nh_grp->nexthops[i];
  2354. if (nh->ifindex == ifindex && nh->nh_weight == weight &&
  2355. ipv6_addr_equal(gw, (struct in6_addr *) nh->gw_addr))
  2356. return true;
  2357. }
  2358. return false;
  2359. }
  2360. static bool
  2361. mlxsw_sp_nexthop6_group_cmp(const struct mlxsw_sp_nexthop_group *nh_grp,
  2362. const struct mlxsw_sp_fib6_entry *fib6_entry)
  2363. {
  2364. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  2365. if (nh_grp->count != fib6_entry->nrt6)
  2366. return false;
  2367. list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
  2368. struct in6_addr *gw;
  2369. int ifindex, weight;
  2370. ifindex = mlxsw_sp_rt6->rt->fib6_nh.nh_dev->ifindex;
  2371. weight = mlxsw_sp_rt6->rt->fib6_nh.nh_weight;
  2372. gw = &mlxsw_sp_rt6->rt->fib6_nh.nh_gw;
  2373. if (!mlxsw_sp_nexthop6_group_has_nexthop(nh_grp, gw, ifindex,
  2374. weight))
  2375. return false;
  2376. }
  2377. return true;
  2378. }
  2379. static int
  2380. mlxsw_sp_nexthop_group_cmp(struct rhashtable_compare_arg *arg, const void *ptr)
  2381. {
  2382. const struct mlxsw_sp_nexthop_group_cmp_arg *cmp_arg = arg->key;
  2383. const struct mlxsw_sp_nexthop_group *nh_grp = ptr;
  2384. switch (cmp_arg->proto) {
  2385. case MLXSW_SP_L3_PROTO_IPV4:
  2386. return cmp_arg->fi != mlxsw_sp_nexthop4_group_fi(nh_grp);
  2387. case MLXSW_SP_L3_PROTO_IPV6:
  2388. return !mlxsw_sp_nexthop6_group_cmp(nh_grp,
  2389. cmp_arg->fib6_entry);
  2390. default:
  2391. WARN_ON(1);
  2392. return 1;
  2393. }
  2394. }
  2395. static int
  2396. mlxsw_sp_nexthop_group_type(const struct mlxsw_sp_nexthop_group *nh_grp)
  2397. {
  2398. return nh_grp->neigh_tbl->family;
  2399. }
  2400. static u32 mlxsw_sp_nexthop_group_hash_obj(const void *data, u32 len, u32 seed)
  2401. {
  2402. const struct mlxsw_sp_nexthop_group *nh_grp = data;
  2403. const struct mlxsw_sp_nexthop *nh;
  2404. struct fib_info *fi;
  2405. unsigned int val;
  2406. int i;
  2407. switch (mlxsw_sp_nexthop_group_type(nh_grp)) {
  2408. case AF_INET:
  2409. fi = mlxsw_sp_nexthop4_group_fi(nh_grp);
  2410. return jhash(&fi, sizeof(fi), seed);
  2411. case AF_INET6:
  2412. val = nh_grp->count;
  2413. for (i = 0; i < nh_grp->count; i++) {
  2414. nh = &nh_grp->nexthops[i];
  2415. val ^= nh->ifindex;
  2416. }
  2417. return jhash(&val, sizeof(val), seed);
  2418. default:
  2419. WARN_ON(1);
  2420. return 0;
  2421. }
  2422. }
  2423. static u32
  2424. mlxsw_sp_nexthop6_group_hash(struct mlxsw_sp_fib6_entry *fib6_entry, u32 seed)
  2425. {
  2426. unsigned int val = fib6_entry->nrt6;
  2427. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  2428. struct net_device *dev;
  2429. list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
  2430. dev = mlxsw_sp_rt6->rt->fib6_nh.nh_dev;
  2431. val ^= dev->ifindex;
  2432. }
  2433. return jhash(&val, sizeof(val), seed);
  2434. }
  2435. static u32
  2436. mlxsw_sp_nexthop_group_hash(const void *data, u32 len, u32 seed)
  2437. {
  2438. const struct mlxsw_sp_nexthop_group_cmp_arg *cmp_arg = data;
  2439. switch (cmp_arg->proto) {
  2440. case MLXSW_SP_L3_PROTO_IPV4:
  2441. return jhash(&cmp_arg->fi, sizeof(cmp_arg->fi), seed);
  2442. case MLXSW_SP_L3_PROTO_IPV6:
  2443. return mlxsw_sp_nexthop6_group_hash(cmp_arg->fib6_entry, seed);
  2444. default:
  2445. WARN_ON(1);
  2446. return 0;
  2447. }
  2448. }
  2449. static const struct rhashtable_params mlxsw_sp_nexthop_group_ht_params = {
  2450. .head_offset = offsetof(struct mlxsw_sp_nexthop_group, ht_node),
  2451. .hashfn = mlxsw_sp_nexthop_group_hash,
  2452. .obj_hashfn = mlxsw_sp_nexthop_group_hash_obj,
  2453. .obj_cmpfn = mlxsw_sp_nexthop_group_cmp,
  2454. };
  2455. static int mlxsw_sp_nexthop_group_insert(struct mlxsw_sp *mlxsw_sp,
  2456. struct mlxsw_sp_nexthop_group *nh_grp)
  2457. {
  2458. if (mlxsw_sp_nexthop_group_type(nh_grp) == AF_INET6 &&
  2459. !nh_grp->gateway)
  2460. return 0;
  2461. return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_group_ht,
  2462. &nh_grp->ht_node,
  2463. mlxsw_sp_nexthop_group_ht_params);
  2464. }
  2465. static void mlxsw_sp_nexthop_group_remove(struct mlxsw_sp *mlxsw_sp,
  2466. struct mlxsw_sp_nexthop_group *nh_grp)
  2467. {
  2468. if (mlxsw_sp_nexthop_group_type(nh_grp) == AF_INET6 &&
  2469. !nh_grp->gateway)
  2470. return;
  2471. rhashtable_remove_fast(&mlxsw_sp->router->nexthop_group_ht,
  2472. &nh_grp->ht_node,
  2473. mlxsw_sp_nexthop_group_ht_params);
  2474. }
  2475. static struct mlxsw_sp_nexthop_group *
  2476. mlxsw_sp_nexthop4_group_lookup(struct mlxsw_sp *mlxsw_sp,
  2477. struct fib_info *fi)
  2478. {
  2479. struct mlxsw_sp_nexthop_group_cmp_arg cmp_arg;
  2480. cmp_arg.proto = MLXSW_SP_L3_PROTO_IPV4;
  2481. cmp_arg.fi = fi;
  2482. return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht,
  2483. &cmp_arg,
  2484. mlxsw_sp_nexthop_group_ht_params);
  2485. }
  2486. static struct mlxsw_sp_nexthop_group *
  2487. mlxsw_sp_nexthop6_group_lookup(struct mlxsw_sp *mlxsw_sp,
  2488. struct mlxsw_sp_fib6_entry *fib6_entry)
  2489. {
  2490. struct mlxsw_sp_nexthop_group_cmp_arg cmp_arg;
  2491. cmp_arg.proto = MLXSW_SP_L3_PROTO_IPV6;
  2492. cmp_arg.fib6_entry = fib6_entry;
  2493. return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_group_ht,
  2494. &cmp_arg,
  2495. mlxsw_sp_nexthop_group_ht_params);
  2496. }
  2497. static const struct rhashtable_params mlxsw_sp_nexthop_ht_params = {
  2498. .key_offset = offsetof(struct mlxsw_sp_nexthop, key),
  2499. .head_offset = offsetof(struct mlxsw_sp_nexthop, ht_node),
  2500. .key_len = sizeof(struct mlxsw_sp_nexthop_key),
  2501. };
  2502. static int mlxsw_sp_nexthop_insert(struct mlxsw_sp *mlxsw_sp,
  2503. struct mlxsw_sp_nexthop *nh)
  2504. {
  2505. return rhashtable_insert_fast(&mlxsw_sp->router->nexthop_ht,
  2506. &nh->ht_node, mlxsw_sp_nexthop_ht_params);
  2507. }
  2508. static void mlxsw_sp_nexthop_remove(struct mlxsw_sp *mlxsw_sp,
  2509. struct mlxsw_sp_nexthop *nh)
  2510. {
  2511. rhashtable_remove_fast(&mlxsw_sp->router->nexthop_ht, &nh->ht_node,
  2512. mlxsw_sp_nexthop_ht_params);
  2513. }
  2514. static struct mlxsw_sp_nexthop *
  2515. mlxsw_sp_nexthop_lookup(struct mlxsw_sp *mlxsw_sp,
  2516. struct mlxsw_sp_nexthop_key key)
  2517. {
  2518. return rhashtable_lookup_fast(&mlxsw_sp->router->nexthop_ht, &key,
  2519. mlxsw_sp_nexthop_ht_params);
  2520. }
  2521. static int mlxsw_sp_adj_index_mass_update_vr(struct mlxsw_sp *mlxsw_sp,
  2522. const struct mlxsw_sp_fib *fib,
  2523. u32 adj_index, u16 ecmp_size,
  2524. u32 new_adj_index,
  2525. u16 new_ecmp_size)
  2526. {
  2527. char raleu_pl[MLXSW_REG_RALEU_LEN];
  2528. mlxsw_reg_raleu_pack(raleu_pl,
  2529. (enum mlxsw_reg_ralxx_protocol) fib->proto,
  2530. fib->vr->id, adj_index, ecmp_size, new_adj_index,
  2531. new_ecmp_size);
  2532. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raleu), raleu_pl);
  2533. }
  2534. static int mlxsw_sp_adj_index_mass_update(struct mlxsw_sp *mlxsw_sp,
  2535. struct mlxsw_sp_nexthop_group *nh_grp,
  2536. u32 old_adj_index, u16 old_ecmp_size)
  2537. {
  2538. struct mlxsw_sp_fib_entry *fib_entry;
  2539. struct mlxsw_sp_fib *fib = NULL;
  2540. int err;
  2541. list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
  2542. if (fib == fib_entry->fib_node->fib)
  2543. continue;
  2544. fib = fib_entry->fib_node->fib;
  2545. err = mlxsw_sp_adj_index_mass_update_vr(mlxsw_sp, fib,
  2546. old_adj_index,
  2547. old_ecmp_size,
  2548. nh_grp->adj_index,
  2549. nh_grp->ecmp_size);
  2550. if (err)
  2551. return err;
  2552. }
  2553. return 0;
  2554. }
  2555. static int __mlxsw_sp_nexthop_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
  2556. struct mlxsw_sp_nexthop *nh)
  2557. {
  2558. struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
  2559. char ratr_pl[MLXSW_REG_RATR_LEN];
  2560. mlxsw_reg_ratr_pack(ratr_pl, MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY,
  2561. true, MLXSW_REG_RATR_TYPE_ETHERNET,
  2562. adj_index, neigh_entry->rif);
  2563. mlxsw_reg_ratr_eth_entry_pack(ratr_pl, neigh_entry->ha);
  2564. if (nh->counter_valid)
  2565. mlxsw_reg_ratr_counter_pack(ratr_pl, nh->counter_index, true);
  2566. else
  2567. mlxsw_reg_ratr_counter_pack(ratr_pl, 0, false);
  2568. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ratr), ratr_pl);
  2569. }
  2570. int mlxsw_sp_nexthop_update(struct mlxsw_sp *mlxsw_sp, u32 adj_index,
  2571. struct mlxsw_sp_nexthop *nh)
  2572. {
  2573. int i;
  2574. for (i = 0; i < nh->num_adj_entries; i++) {
  2575. int err;
  2576. err = __mlxsw_sp_nexthop_update(mlxsw_sp, adj_index + i, nh);
  2577. if (err)
  2578. return err;
  2579. }
  2580. return 0;
  2581. }
  2582. static int __mlxsw_sp_nexthop_ipip_update(struct mlxsw_sp *mlxsw_sp,
  2583. u32 adj_index,
  2584. struct mlxsw_sp_nexthop *nh)
  2585. {
  2586. const struct mlxsw_sp_ipip_ops *ipip_ops;
  2587. ipip_ops = mlxsw_sp->router->ipip_ops_arr[nh->ipip_entry->ipipt];
  2588. return ipip_ops->nexthop_update(mlxsw_sp, adj_index, nh->ipip_entry);
  2589. }
  2590. static int mlxsw_sp_nexthop_ipip_update(struct mlxsw_sp *mlxsw_sp,
  2591. u32 adj_index,
  2592. struct mlxsw_sp_nexthop *nh)
  2593. {
  2594. int i;
  2595. for (i = 0; i < nh->num_adj_entries; i++) {
  2596. int err;
  2597. err = __mlxsw_sp_nexthop_ipip_update(mlxsw_sp, adj_index + i,
  2598. nh);
  2599. if (err)
  2600. return err;
  2601. }
  2602. return 0;
  2603. }
  2604. static int
  2605. mlxsw_sp_nexthop_group_update(struct mlxsw_sp *mlxsw_sp,
  2606. struct mlxsw_sp_nexthop_group *nh_grp,
  2607. bool reallocate)
  2608. {
  2609. u32 adj_index = nh_grp->adj_index; /* base */
  2610. struct mlxsw_sp_nexthop *nh;
  2611. int i;
  2612. int err;
  2613. for (i = 0; i < nh_grp->count; i++) {
  2614. nh = &nh_grp->nexthops[i];
  2615. if (!nh->should_offload) {
  2616. nh->offloaded = 0;
  2617. continue;
  2618. }
  2619. if (nh->update || reallocate) {
  2620. switch (nh->type) {
  2621. case MLXSW_SP_NEXTHOP_TYPE_ETH:
  2622. err = mlxsw_sp_nexthop_update
  2623. (mlxsw_sp, adj_index, nh);
  2624. break;
  2625. case MLXSW_SP_NEXTHOP_TYPE_IPIP:
  2626. err = mlxsw_sp_nexthop_ipip_update
  2627. (mlxsw_sp, adj_index, nh);
  2628. break;
  2629. }
  2630. if (err)
  2631. return err;
  2632. nh->update = 0;
  2633. nh->offloaded = 1;
  2634. }
  2635. adj_index += nh->num_adj_entries;
  2636. }
  2637. return 0;
  2638. }
  2639. static bool
  2640. mlxsw_sp_fib_node_entry_is_first(const struct mlxsw_sp_fib_node *fib_node,
  2641. const struct mlxsw_sp_fib_entry *fib_entry);
  2642. static int
  2643. mlxsw_sp_nexthop_fib_entries_update(struct mlxsw_sp *mlxsw_sp,
  2644. struct mlxsw_sp_nexthop_group *nh_grp)
  2645. {
  2646. struct mlxsw_sp_fib_entry *fib_entry;
  2647. int err;
  2648. list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
  2649. if (!mlxsw_sp_fib_node_entry_is_first(fib_entry->fib_node,
  2650. fib_entry))
  2651. continue;
  2652. err = mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
  2653. if (err)
  2654. return err;
  2655. }
  2656. return 0;
  2657. }
  2658. static void
  2659. mlxsw_sp_fib_entry_offload_refresh(struct mlxsw_sp_fib_entry *fib_entry,
  2660. enum mlxsw_reg_ralue_op op, int err);
  2661. static void
  2662. mlxsw_sp_nexthop_fib_entries_refresh(struct mlxsw_sp_nexthop_group *nh_grp)
  2663. {
  2664. enum mlxsw_reg_ralue_op op = MLXSW_REG_RALUE_OP_WRITE_WRITE;
  2665. struct mlxsw_sp_fib_entry *fib_entry;
  2666. list_for_each_entry(fib_entry, &nh_grp->fib_list, nexthop_group_node) {
  2667. if (!mlxsw_sp_fib_node_entry_is_first(fib_entry->fib_node,
  2668. fib_entry))
  2669. continue;
  2670. mlxsw_sp_fib_entry_offload_refresh(fib_entry, op, 0);
  2671. }
  2672. }
  2673. static void mlxsw_sp_adj_grp_size_round_up(u16 *p_adj_grp_size)
  2674. {
  2675. /* Valid sizes for an adjacency group are:
  2676. * 1-64, 512, 1024, 2048 and 4096.
  2677. */
  2678. if (*p_adj_grp_size <= 64)
  2679. return;
  2680. else if (*p_adj_grp_size <= 512)
  2681. *p_adj_grp_size = 512;
  2682. else if (*p_adj_grp_size <= 1024)
  2683. *p_adj_grp_size = 1024;
  2684. else if (*p_adj_grp_size <= 2048)
  2685. *p_adj_grp_size = 2048;
  2686. else
  2687. *p_adj_grp_size = 4096;
  2688. }
  2689. static void mlxsw_sp_adj_grp_size_round_down(u16 *p_adj_grp_size,
  2690. unsigned int alloc_size)
  2691. {
  2692. if (alloc_size >= 4096)
  2693. *p_adj_grp_size = 4096;
  2694. else if (alloc_size >= 2048)
  2695. *p_adj_grp_size = 2048;
  2696. else if (alloc_size >= 1024)
  2697. *p_adj_grp_size = 1024;
  2698. else if (alloc_size >= 512)
  2699. *p_adj_grp_size = 512;
  2700. }
  2701. static int mlxsw_sp_fix_adj_grp_size(struct mlxsw_sp *mlxsw_sp,
  2702. u16 *p_adj_grp_size)
  2703. {
  2704. unsigned int alloc_size;
  2705. int err;
  2706. /* Round up the requested group size to the next size supported
  2707. * by the device and make sure the request can be satisfied.
  2708. */
  2709. mlxsw_sp_adj_grp_size_round_up(p_adj_grp_size);
  2710. err = mlxsw_sp_kvdl_alloc_size_query(mlxsw_sp, *p_adj_grp_size,
  2711. &alloc_size);
  2712. if (err)
  2713. return err;
  2714. /* It is possible the allocation results in more allocated
  2715. * entries than requested. Try to use as much of them as
  2716. * possible.
  2717. */
  2718. mlxsw_sp_adj_grp_size_round_down(p_adj_grp_size, alloc_size);
  2719. return 0;
  2720. }
  2721. static void
  2722. mlxsw_sp_nexthop_group_normalize(struct mlxsw_sp_nexthop_group *nh_grp)
  2723. {
  2724. int i, g = 0, sum_norm_weight = 0;
  2725. struct mlxsw_sp_nexthop *nh;
  2726. for (i = 0; i < nh_grp->count; i++) {
  2727. nh = &nh_grp->nexthops[i];
  2728. if (!nh->should_offload)
  2729. continue;
  2730. if (g > 0)
  2731. g = gcd(nh->nh_weight, g);
  2732. else
  2733. g = nh->nh_weight;
  2734. }
  2735. for (i = 0; i < nh_grp->count; i++) {
  2736. nh = &nh_grp->nexthops[i];
  2737. if (!nh->should_offload)
  2738. continue;
  2739. nh->norm_nh_weight = nh->nh_weight / g;
  2740. sum_norm_weight += nh->norm_nh_weight;
  2741. }
  2742. nh_grp->sum_norm_weight = sum_norm_weight;
  2743. }
  2744. static void
  2745. mlxsw_sp_nexthop_group_rebalance(struct mlxsw_sp_nexthop_group *nh_grp)
  2746. {
  2747. int total = nh_grp->sum_norm_weight;
  2748. u16 ecmp_size = nh_grp->ecmp_size;
  2749. int i, weight = 0, lower_bound = 0;
  2750. for (i = 0; i < nh_grp->count; i++) {
  2751. struct mlxsw_sp_nexthop *nh = &nh_grp->nexthops[i];
  2752. int upper_bound;
  2753. if (!nh->should_offload)
  2754. continue;
  2755. weight += nh->norm_nh_weight;
  2756. upper_bound = DIV_ROUND_CLOSEST(ecmp_size * weight, total);
  2757. nh->num_adj_entries = upper_bound - lower_bound;
  2758. lower_bound = upper_bound;
  2759. }
  2760. }
  2761. static void
  2762. mlxsw_sp_nexthop_group_refresh(struct mlxsw_sp *mlxsw_sp,
  2763. struct mlxsw_sp_nexthop_group *nh_grp)
  2764. {
  2765. u16 ecmp_size, old_ecmp_size;
  2766. struct mlxsw_sp_nexthop *nh;
  2767. bool offload_change = false;
  2768. u32 adj_index;
  2769. bool old_adj_index_valid;
  2770. u32 old_adj_index;
  2771. int i;
  2772. int err;
  2773. if (!nh_grp->gateway) {
  2774. mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
  2775. return;
  2776. }
  2777. for (i = 0; i < nh_grp->count; i++) {
  2778. nh = &nh_grp->nexthops[i];
  2779. if (nh->should_offload != nh->offloaded) {
  2780. offload_change = true;
  2781. if (nh->should_offload)
  2782. nh->update = 1;
  2783. }
  2784. }
  2785. if (!offload_change) {
  2786. /* Nothing was added or removed, so no need to reallocate. Just
  2787. * update MAC on existing adjacency indexes.
  2788. */
  2789. err = mlxsw_sp_nexthop_group_update(mlxsw_sp, nh_grp, false);
  2790. if (err) {
  2791. dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
  2792. goto set_trap;
  2793. }
  2794. return;
  2795. }
  2796. mlxsw_sp_nexthop_group_normalize(nh_grp);
  2797. if (!nh_grp->sum_norm_weight)
  2798. /* No neigh of this group is connected so we just set
  2799. * the trap and let everthing flow through kernel.
  2800. */
  2801. goto set_trap;
  2802. ecmp_size = nh_grp->sum_norm_weight;
  2803. err = mlxsw_sp_fix_adj_grp_size(mlxsw_sp, &ecmp_size);
  2804. if (err)
  2805. /* No valid allocation size available. */
  2806. goto set_trap;
  2807. err = mlxsw_sp_kvdl_alloc(mlxsw_sp, ecmp_size, &adj_index);
  2808. if (err) {
  2809. /* We ran out of KVD linear space, just set the
  2810. * trap and let everything flow through kernel.
  2811. */
  2812. dev_warn(mlxsw_sp->bus_info->dev, "Failed to allocate KVD linear area for nexthop group.\n");
  2813. goto set_trap;
  2814. }
  2815. old_adj_index_valid = nh_grp->adj_index_valid;
  2816. old_adj_index = nh_grp->adj_index;
  2817. old_ecmp_size = nh_grp->ecmp_size;
  2818. nh_grp->adj_index_valid = 1;
  2819. nh_grp->adj_index = adj_index;
  2820. nh_grp->ecmp_size = ecmp_size;
  2821. mlxsw_sp_nexthop_group_rebalance(nh_grp);
  2822. err = mlxsw_sp_nexthop_group_update(mlxsw_sp, nh_grp, true);
  2823. if (err) {
  2824. dev_warn(mlxsw_sp->bus_info->dev, "Failed to update neigh MAC in adjacency table.\n");
  2825. goto set_trap;
  2826. }
  2827. if (!old_adj_index_valid) {
  2828. /* The trap was set for fib entries, so we have to call
  2829. * fib entry update to unset it and use adjacency index.
  2830. */
  2831. err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
  2832. if (err) {
  2833. dev_warn(mlxsw_sp->bus_info->dev, "Failed to add adjacency index to fib entries.\n");
  2834. goto set_trap;
  2835. }
  2836. return;
  2837. }
  2838. err = mlxsw_sp_adj_index_mass_update(mlxsw_sp, nh_grp,
  2839. old_adj_index, old_ecmp_size);
  2840. mlxsw_sp_kvdl_free(mlxsw_sp, old_adj_index);
  2841. if (err) {
  2842. dev_warn(mlxsw_sp->bus_info->dev, "Failed to mass-update adjacency index for nexthop group.\n");
  2843. goto set_trap;
  2844. }
  2845. /* Offload state within the group changed, so update the flags. */
  2846. mlxsw_sp_nexthop_fib_entries_refresh(nh_grp);
  2847. return;
  2848. set_trap:
  2849. old_adj_index_valid = nh_grp->adj_index_valid;
  2850. nh_grp->adj_index_valid = 0;
  2851. for (i = 0; i < nh_grp->count; i++) {
  2852. nh = &nh_grp->nexthops[i];
  2853. nh->offloaded = 0;
  2854. }
  2855. err = mlxsw_sp_nexthop_fib_entries_update(mlxsw_sp, nh_grp);
  2856. if (err)
  2857. dev_warn(mlxsw_sp->bus_info->dev, "Failed to set traps for fib entries.\n");
  2858. if (old_adj_index_valid)
  2859. mlxsw_sp_kvdl_free(mlxsw_sp, nh_grp->adj_index);
  2860. }
  2861. static void __mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp_nexthop *nh,
  2862. bool removing)
  2863. {
  2864. if (!removing)
  2865. nh->should_offload = 1;
  2866. else
  2867. nh->should_offload = 0;
  2868. nh->update = 1;
  2869. }
  2870. static void
  2871. mlxsw_sp_nexthop_neigh_update(struct mlxsw_sp *mlxsw_sp,
  2872. struct mlxsw_sp_neigh_entry *neigh_entry,
  2873. bool removing)
  2874. {
  2875. struct mlxsw_sp_nexthop *nh;
  2876. list_for_each_entry(nh, &neigh_entry->nexthop_list,
  2877. neigh_list_node) {
  2878. __mlxsw_sp_nexthop_neigh_update(nh, removing);
  2879. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
  2880. }
  2881. }
  2882. static void mlxsw_sp_nexthop_rif_init(struct mlxsw_sp_nexthop *nh,
  2883. struct mlxsw_sp_rif *rif)
  2884. {
  2885. if (nh->rif)
  2886. return;
  2887. nh->rif = rif;
  2888. list_add(&nh->rif_list_node, &rif->nexthop_list);
  2889. }
  2890. static void mlxsw_sp_nexthop_rif_fini(struct mlxsw_sp_nexthop *nh)
  2891. {
  2892. if (!nh->rif)
  2893. return;
  2894. list_del(&nh->rif_list_node);
  2895. nh->rif = NULL;
  2896. }
  2897. static int mlxsw_sp_nexthop_neigh_init(struct mlxsw_sp *mlxsw_sp,
  2898. struct mlxsw_sp_nexthop *nh)
  2899. {
  2900. struct mlxsw_sp_neigh_entry *neigh_entry;
  2901. struct neighbour *n;
  2902. u8 nud_state, dead;
  2903. int err;
  2904. if (!nh->nh_grp->gateway || nh->neigh_entry)
  2905. return 0;
  2906. /* Take a reference of neigh here ensuring that neigh would
  2907. * not be destructed before the nexthop entry is finished.
  2908. * The reference is taken either in neigh_lookup() or
  2909. * in neigh_create() in case n is not found.
  2910. */
  2911. n = neigh_lookup(nh->nh_grp->neigh_tbl, &nh->gw_addr, nh->rif->dev);
  2912. if (!n) {
  2913. n = neigh_create(nh->nh_grp->neigh_tbl, &nh->gw_addr,
  2914. nh->rif->dev);
  2915. if (IS_ERR(n))
  2916. return PTR_ERR(n);
  2917. neigh_event_send(n, NULL);
  2918. }
  2919. neigh_entry = mlxsw_sp_neigh_entry_lookup(mlxsw_sp, n);
  2920. if (!neigh_entry) {
  2921. neigh_entry = mlxsw_sp_neigh_entry_create(mlxsw_sp, n);
  2922. if (IS_ERR(neigh_entry)) {
  2923. err = -EINVAL;
  2924. goto err_neigh_entry_create;
  2925. }
  2926. }
  2927. /* If that is the first nexthop connected to that neigh, add to
  2928. * nexthop_neighs_list
  2929. */
  2930. if (list_empty(&neigh_entry->nexthop_list))
  2931. list_add_tail(&neigh_entry->nexthop_neighs_list_node,
  2932. &mlxsw_sp->router->nexthop_neighs_list);
  2933. nh->neigh_entry = neigh_entry;
  2934. list_add_tail(&nh->neigh_list_node, &neigh_entry->nexthop_list);
  2935. read_lock_bh(&n->lock);
  2936. nud_state = n->nud_state;
  2937. dead = n->dead;
  2938. read_unlock_bh(&n->lock);
  2939. __mlxsw_sp_nexthop_neigh_update(nh, !(nud_state & NUD_VALID && !dead));
  2940. return 0;
  2941. err_neigh_entry_create:
  2942. neigh_release(n);
  2943. return err;
  2944. }
  2945. static void mlxsw_sp_nexthop_neigh_fini(struct mlxsw_sp *mlxsw_sp,
  2946. struct mlxsw_sp_nexthop *nh)
  2947. {
  2948. struct mlxsw_sp_neigh_entry *neigh_entry = nh->neigh_entry;
  2949. struct neighbour *n;
  2950. if (!neigh_entry)
  2951. return;
  2952. n = neigh_entry->key.n;
  2953. __mlxsw_sp_nexthop_neigh_update(nh, true);
  2954. list_del(&nh->neigh_list_node);
  2955. nh->neigh_entry = NULL;
  2956. /* If that is the last nexthop connected to that neigh, remove from
  2957. * nexthop_neighs_list
  2958. */
  2959. if (list_empty(&neigh_entry->nexthop_list))
  2960. list_del(&neigh_entry->nexthop_neighs_list_node);
  2961. if (!neigh_entry->connected && list_empty(&neigh_entry->nexthop_list))
  2962. mlxsw_sp_neigh_entry_destroy(mlxsw_sp, neigh_entry);
  2963. neigh_release(n);
  2964. }
  2965. static bool mlxsw_sp_ipip_netdev_ul_up(struct net_device *ol_dev)
  2966. {
  2967. struct net_device *ul_dev = __mlxsw_sp_ipip_netdev_ul_dev_get(ol_dev);
  2968. return ul_dev ? (ul_dev->flags & IFF_UP) : true;
  2969. }
  2970. static void mlxsw_sp_nexthop_ipip_init(struct mlxsw_sp *mlxsw_sp,
  2971. struct mlxsw_sp_nexthop *nh,
  2972. struct mlxsw_sp_ipip_entry *ipip_entry)
  2973. {
  2974. bool removing;
  2975. if (!nh->nh_grp->gateway || nh->ipip_entry)
  2976. return;
  2977. nh->ipip_entry = ipip_entry;
  2978. removing = !mlxsw_sp_ipip_netdev_ul_up(ipip_entry->ol_dev);
  2979. __mlxsw_sp_nexthop_neigh_update(nh, removing);
  2980. mlxsw_sp_nexthop_rif_init(nh, &ipip_entry->ol_lb->common);
  2981. }
  2982. static void mlxsw_sp_nexthop_ipip_fini(struct mlxsw_sp *mlxsw_sp,
  2983. struct mlxsw_sp_nexthop *nh)
  2984. {
  2985. struct mlxsw_sp_ipip_entry *ipip_entry = nh->ipip_entry;
  2986. if (!ipip_entry)
  2987. return;
  2988. __mlxsw_sp_nexthop_neigh_update(nh, true);
  2989. nh->ipip_entry = NULL;
  2990. }
  2991. static bool mlxsw_sp_nexthop4_ipip_type(const struct mlxsw_sp *mlxsw_sp,
  2992. const struct fib_nh *fib_nh,
  2993. enum mlxsw_sp_ipip_type *p_ipipt)
  2994. {
  2995. struct net_device *dev = fib_nh->nh_dev;
  2996. return dev &&
  2997. fib_nh->nh_parent->fib_type == RTN_UNICAST &&
  2998. mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, p_ipipt);
  2999. }
  3000. static void mlxsw_sp_nexthop_type_fini(struct mlxsw_sp *mlxsw_sp,
  3001. struct mlxsw_sp_nexthop *nh)
  3002. {
  3003. switch (nh->type) {
  3004. case MLXSW_SP_NEXTHOP_TYPE_ETH:
  3005. mlxsw_sp_nexthop_neigh_fini(mlxsw_sp, nh);
  3006. mlxsw_sp_nexthop_rif_fini(nh);
  3007. break;
  3008. case MLXSW_SP_NEXTHOP_TYPE_IPIP:
  3009. mlxsw_sp_nexthop_rif_fini(nh);
  3010. mlxsw_sp_nexthop_ipip_fini(mlxsw_sp, nh);
  3011. break;
  3012. }
  3013. }
  3014. static int mlxsw_sp_nexthop4_type_init(struct mlxsw_sp *mlxsw_sp,
  3015. struct mlxsw_sp_nexthop *nh,
  3016. struct fib_nh *fib_nh)
  3017. {
  3018. const struct mlxsw_sp_ipip_ops *ipip_ops;
  3019. struct net_device *dev = fib_nh->nh_dev;
  3020. struct mlxsw_sp_ipip_entry *ipip_entry;
  3021. struct mlxsw_sp_rif *rif;
  3022. int err;
  3023. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, dev);
  3024. if (ipip_entry) {
  3025. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
  3026. if (ipip_ops->can_offload(mlxsw_sp, dev,
  3027. MLXSW_SP_L3_PROTO_IPV4)) {
  3028. nh->type = MLXSW_SP_NEXTHOP_TYPE_IPIP;
  3029. mlxsw_sp_nexthop_ipip_init(mlxsw_sp, nh, ipip_entry);
  3030. return 0;
  3031. }
  3032. }
  3033. nh->type = MLXSW_SP_NEXTHOP_TYPE_ETH;
  3034. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  3035. if (!rif)
  3036. return 0;
  3037. mlxsw_sp_nexthop_rif_init(nh, rif);
  3038. err = mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
  3039. if (err)
  3040. goto err_neigh_init;
  3041. return 0;
  3042. err_neigh_init:
  3043. mlxsw_sp_nexthop_rif_fini(nh);
  3044. return err;
  3045. }
  3046. static void mlxsw_sp_nexthop4_type_fini(struct mlxsw_sp *mlxsw_sp,
  3047. struct mlxsw_sp_nexthop *nh)
  3048. {
  3049. mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
  3050. }
  3051. static int mlxsw_sp_nexthop4_init(struct mlxsw_sp *mlxsw_sp,
  3052. struct mlxsw_sp_nexthop_group *nh_grp,
  3053. struct mlxsw_sp_nexthop *nh,
  3054. struct fib_nh *fib_nh)
  3055. {
  3056. struct net_device *dev = fib_nh->nh_dev;
  3057. struct in_device *in_dev;
  3058. int err;
  3059. nh->nh_grp = nh_grp;
  3060. nh->key.fib_nh = fib_nh;
  3061. #ifdef CONFIG_IP_ROUTE_MULTIPATH
  3062. nh->nh_weight = fib_nh->nh_weight;
  3063. #else
  3064. nh->nh_weight = 1;
  3065. #endif
  3066. memcpy(&nh->gw_addr, &fib_nh->nh_gw, sizeof(fib_nh->nh_gw));
  3067. err = mlxsw_sp_nexthop_insert(mlxsw_sp, nh);
  3068. if (err)
  3069. return err;
  3070. mlxsw_sp_nexthop_counter_alloc(mlxsw_sp, nh);
  3071. list_add_tail(&nh->router_list_node, &mlxsw_sp->router->nexthop_list);
  3072. if (!dev)
  3073. return 0;
  3074. in_dev = __in_dev_get_rtnl(dev);
  3075. if (in_dev && IN_DEV_IGNORE_ROUTES_WITH_LINKDOWN(in_dev) &&
  3076. fib_nh->nh_flags & RTNH_F_LINKDOWN)
  3077. return 0;
  3078. err = mlxsw_sp_nexthop4_type_init(mlxsw_sp, nh, fib_nh);
  3079. if (err)
  3080. goto err_nexthop_neigh_init;
  3081. return 0;
  3082. err_nexthop_neigh_init:
  3083. mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
  3084. return err;
  3085. }
  3086. static void mlxsw_sp_nexthop4_fini(struct mlxsw_sp *mlxsw_sp,
  3087. struct mlxsw_sp_nexthop *nh)
  3088. {
  3089. mlxsw_sp_nexthop4_type_fini(mlxsw_sp, nh);
  3090. list_del(&nh->router_list_node);
  3091. mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
  3092. mlxsw_sp_nexthop_remove(mlxsw_sp, nh);
  3093. }
  3094. static void mlxsw_sp_nexthop4_event(struct mlxsw_sp *mlxsw_sp,
  3095. unsigned long event, struct fib_nh *fib_nh)
  3096. {
  3097. struct mlxsw_sp_nexthop_key key;
  3098. struct mlxsw_sp_nexthop *nh;
  3099. if (mlxsw_sp->router->aborted)
  3100. return;
  3101. key.fib_nh = fib_nh;
  3102. nh = mlxsw_sp_nexthop_lookup(mlxsw_sp, key);
  3103. if (WARN_ON_ONCE(!nh))
  3104. return;
  3105. switch (event) {
  3106. case FIB_EVENT_NH_ADD:
  3107. mlxsw_sp_nexthop4_type_init(mlxsw_sp, nh, fib_nh);
  3108. break;
  3109. case FIB_EVENT_NH_DEL:
  3110. mlxsw_sp_nexthop4_type_fini(mlxsw_sp, nh);
  3111. break;
  3112. }
  3113. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
  3114. }
  3115. static void mlxsw_sp_nexthop_rif_update(struct mlxsw_sp *mlxsw_sp,
  3116. struct mlxsw_sp_rif *rif)
  3117. {
  3118. struct mlxsw_sp_nexthop *nh;
  3119. bool removing;
  3120. list_for_each_entry(nh, &rif->nexthop_list, rif_list_node) {
  3121. switch (nh->type) {
  3122. case MLXSW_SP_NEXTHOP_TYPE_ETH:
  3123. removing = false;
  3124. break;
  3125. case MLXSW_SP_NEXTHOP_TYPE_IPIP:
  3126. removing = !mlxsw_sp_ipip_netdev_ul_up(rif->dev);
  3127. break;
  3128. default:
  3129. WARN_ON(1);
  3130. continue;
  3131. }
  3132. __mlxsw_sp_nexthop_neigh_update(nh, removing);
  3133. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
  3134. }
  3135. }
  3136. static void mlxsw_sp_nexthop_rif_migrate(struct mlxsw_sp *mlxsw_sp,
  3137. struct mlxsw_sp_rif *old_rif,
  3138. struct mlxsw_sp_rif *new_rif)
  3139. {
  3140. struct mlxsw_sp_nexthop *nh;
  3141. list_splice_init(&old_rif->nexthop_list, &new_rif->nexthop_list);
  3142. list_for_each_entry(nh, &new_rif->nexthop_list, rif_list_node)
  3143. nh->rif = new_rif;
  3144. mlxsw_sp_nexthop_rif_update(mlxsw_sp, new_rif);
  3145. }
  3146. static void mlxsw_sp_nexthop_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
  3147. struct mlxsw_sp_rif *rif)
  3148. {
  3149. struct mlxsw_sp_nexthop *nh, *tmp;
  3150. list_for_each_entry_safe(nh, tmp, &rif->nexthop_list, rif_list_node) {
  3151. mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
  3152. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh->nh_grp);
  3153. }
  3154. }
  3155. static bool mlxsw_sp_fi_is_gateway(const struct mlxsw_sp *mlxsw_sp,
  3156. const struct fib_info *fi)
  3157. {
  3158. return fi->fib_nh->nh_scope == RT_SCOPE_LINK ||
  3159. mlxsw_sp_nexthop4_ipip_type(mlxsw_sp, fi->fib_nh, NULL);
  3160. }
  3161. static struct mlxsw_sp_nexthop_group *
  3162. mlxsw_sp_nexthop4_group_create(struct mlxsw_sp *mlxsw_sp, struct fib_info *fi)
  3163. {
  3164. struct mlxsw_sp_nexthop_group *nh_grp;
  3165. struct mlxsw_sp_nexthop *nh;
  3166. struct fib_nh *fib_nh;
  3167. size_t alloc_size;
  3168. int i;
  3169. int err;
  3170. alloc_size = sizeof(*nh_grp) +
  3171. fi->fib_nhs * sizeof(struct mlxsw_sp_nexthop);
  3172. nh_grp = kzalloc(alloc_size, GFP_KERNEL);
  3173. if (!nh_grp)
  3174. return ERR_PTR(-ENOMEM);
  3175. nh_grp->priv = fi;
  3176. INIT_LIST_HEAD(&nh_grp->fib_list);
  3177. nh_grp->neigh_tbl = &arp_tbl;
  3178. nh_grp->gateway = mlxsw_sp_fi_is_gateway(mlxsw_sp, fi);
  3179. nh_grp->count = fi->fib_nhs;
  3180. fib_info_hold(fi);
  3181. for (i = 0; i < nh_grp->count; i++) {
  3182. nh = &nh_grp->nexthops[i];
  3183. fib_nh = &fi->fib_nh[i];
  3184. err = mlxsw_sp_nexthop4_init(mlxsw_sp, nh_grp, nh, fib_nh);
  3185. if (err)
  3186. goto err_nexthop4_init;
  3187. }
  3188. err = mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
  3189. if (err)
  3190. goto err_nexthop_group_insert;
  3191. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
  3192. return nh_grp;
  3193. err_nexthop_group_insert:
  3194. err_nexthop4_init:
  3195. for (i--; i >= 0; i--) {
  3196. nh = &nh_grp->nexthops[i];
  3197. mlxsw_sp_nexthop4_fini(mlxsw_sp, nh);
  3198. }
  3199. fib_info_put(fi);
  3200. kfree(nh_grp);
  3201. return ERR_PTR(err);
  3202. }
  3203. static void
  3204. mlxsw_sp_nexthop4_group_destroy(struct mlxsw_sp *mlxsw_sp,
  3205. struct mlxsw_sp_nexthop_group *nh_grp)
  3206. {
  3207. struct mlxsw_sp_nexthop *nh;
  3208. int i;
  3209. mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
  3210. for (i = 0; i < nh_grp->count; i++) {
  3211. nh = &nh_grp->nexthops[i];
  3212. mlxsw_sp_nexthop4_fini(mlxsw_sp, nh);
  3213. }
  3214. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
  3215. WARN_ON_ONCE(nh_grp->adj_index_valid);
  3216. fib_info_put(mlxsw_sp_nexthop4_group_fi(nh_grp));
  3217. kfree(nh_grp);
  3218. }
  3219. static int mlxsw_sp_nexthop4_group_get(struct mlxsw_sp *mlxsw_sp,
  3220. struct mlxsw_sp_fib_entry *fib_entry,
  3221. struct fib_info *fi)
  3222. {
  3223. struct mlxsw_sp_nexthop_group *nh_grp;
  3224. nh_grp = mlxsw_sp_nexthop4_group_lookup(mlxsw_sp, fi);
  3225. if (!nh_grp) {
  3226. nh_grp = mlxsw_sp_nexthop4_group_create(mlxsw_sp, fi);
  3227. if (IS_ERR(nh_grp))
  3228. return PTR_ERR(nh_grp);
  3229. }
  3230. list_add_tail(&fib_entry->nexthop_group_node, &nh_grp->fib_list);
  3231. fib_entry->nh_group = nh_grp;
  3232. return 0;
  3233. }
  3234. static void mlxsw_sp_nexthop4_group_put(struct mlxsw_sp *mlxsw_sp,
  3235. struct mlxsw_sp_fib_entry *fib_entry)
  3236. {
  3237. struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
  3238. list_del(&fib_entry->nexthop_group_node);
  3239. if (!list_empty(&nh_grp->fib_list))
  3240. return;
  3241. mlxsw_sp_nexthop4_group_destroy(mlxsw_sp, nh_grp);
  3242. }
  3243. static bool
  3244. mlxsw_sp_fib4_entry_should_offload(const struct mlxsw_sp_fib_entry *fib_entry)
  3245. {
  3246. struct mlxsw_sp_fib4_entry *fib4_entry;
  3247. fib4_entry = container_of(fib_entry, struct mlxsw_sp_fib4_entry,
  3248. common);
  3249. return !fib4_entry->tos;
  3250. }
  3251. static bool
  3252. mlxsw_sp_fib_entry_should_offload(const struct mlxsw_sp_fib_entry *fib_entry)
  3253. {
  3254. struct mlxsw_sp_nexthop_group *nh_group = fib_entry->nh_group;
  3255. switch (fib_entry->fib_node->fib->proto) {
  3256. case MLXSW_SP_L3_PROTO_IPV4:
  3257. if (!mlxsw_sp_fib4_entry_should_offload(fib_entry))
  3258. return false;
  3259. break;
  3260. case MLXSW_SP_L3_PROTO_IPV6:
  3261. break;
  3262. }
  3263. switch (fib_entry->type) {
  3264. case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
  3265. return !!nh_group->adj_index_valid;
  3266. case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
  3267. return !!nh_group->nh_rif;
  3268. case MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP:
  3269. return true;
  3270. default:
  3271. return false;
  3272. }
  3273. }
  3274. static struct mlxsw_sp_nexthop *
  3275. mlxsw_sp_rt6_nexthop(struct mlxsw_sp_nexthop_group *nh_grp,
  3276. const struct mlxsw_sp_rt6 *mlxsw_sp_rt6)
  3277. {
  3278. int i;
  3279. for (i = 0; i < nh_grp->count; i++) {
  3280. struct mlxsw_sp_nexthop *nh = &nh_grp->nexthops[i];
  3281. struct fib6_info *rt = mlxsw_sp_rt6->rt;
  3282. if (nh->rif && nh->rif->dev == rt->fib6_nh.nh_dev &&
  3283. ipv6_addr_equal((const struct in6_addr *) &nh->gw_addr,
  3284. &rt->fib6_nh.nh_gw))
  3285. return nh;
  3286. continue;
  3287. }
  3288. return NULL;
  3289. }
  3290. static void
  3291. mlxsw_sp_fib4_entry_offload_set(struct mlxsw_sp_fib_entry *fib_entry)
  3292. {
  3293. struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
  3294. int i;
  3295. if (fib_entry->type == MLXSW_SP_FIB_ENTRY_TYPE_LOCAL ||
  3296. fib_entry->type == MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP) {
  3297. nh_grp->nexthops->key.fib_nh->nh_flags |= RTNH_F_OFFLOAD;
  3298. return;
  3299. }
  3300. for (i = 0; i < nh_grp->count; i++) {
  3301. struct mlxsw_sp_nexthop *nh = &nh_grp->nexthops[i];
  3302. if (nh->offloaded)
  3303. nh->key.fib_nh->nh_flags |= RTNH_F_OFFLOAD;
  3304. else
  3305. nh->key.fib_nh->nh_flags &= ~RTNH_F_OFFLOAD;
  3306. }
  3307. }
  3308. static void
  3309. mlxsw_sp_fib4_entry_offload_unset(struct mlxsw_sp_fib_entry *fib_entry)
  3310. {
  3311. struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
  3312. int i;
  3313. if (!list_is_singular(&nh_grp->fib_list))
  3314. return;
  3315. for (i = 0; i < nh_grp->count; i++) {
  3316. struct mlxsw_sp_nexthop *nh = &nh_grp->nexthops[i];
  3317. nh->key.fib_nh->nh_flags &= ~RTNH_F_OFFLOAD;
  3318. }
  3319. }
  3320. static void
  3321. mlxsw_sp_fib6_entry_offload_set(struct mlxsw_sp_fib_entry *fib_entry)
  3322. {
  3323. struct mlxsw_sp_fib6_entry *fib6_entry;
  3324. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  3325. fib6_entry = container_of(fib_entry, struct mlxsw_sp_fib6_entry,
  3326. common);
  3327. if (fib_entry->type == MLXSW_SP_FIB_ENTRY_TYPE_LOCAL) {
  3328. list_first_entry(&fib6_entry->rt6_list, struct mlxsw_sp_rt6,
  3329. list)->rt->fib6_nh.nh_flags |= RTNH_F_OFFLOAD;
  3330. return;
  3331. }
  3332. list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
  3333. struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
  3334. struct mlxsw_sp_nexthop *nh;
  3335. nh = mlxsw_sp_rt6_nexthop(nh_grp, mlxsw_sp_rt6);
  3336. if (nh && nh->offloaded)
  3337. mlxsw_sp_rt6->rt->fib6_nh.nh_flags |= RTNH_F_OFFLOAD;
  3338. else
  3339. mlxsw_sp_rt6->rt->fib6_nh.nh_flags &= ~RTNH_F_OFFLOAD;
  3340. }
  3341. }
  3342. static void
  3343. mlxsw_sp_fib6_entry_offload_unset(struct mlxsw_sp_fib_entry *fib_entry)
  3344. {
  3345. struct mlxsw_sp_fib6_entry *fib6_entry;
  3346. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  3347. fib6_entry = container_of(fib_entry, struct mlxsw_sp_fib6_entry,
  3348. common);
  3349. list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
  3350. struct fib6_info *rt = mlxsw_sp_rt6->rt;
  3351. rt->fib6_nh.nh_flags &= ~RTNH_F_OFFLOAD;
  3352. }
  3353. }
  3354. static void mlxsw_sp_fib_entry_offload_set(struct mlxsw_sp_fib_entry *fib_entry)
  3355. {
  3356. switch (fib_entry->fib_node->fib->proto) {
  3357. case MLXSW_SP_L3_PROTO_IPV4:
  3358. mlxsw_sp_fib4_entry_offload_set(fib_entry);
  3359. break;
  3360. case MLXSW_SP_L3_PROTO_IPV6:
  3361. mlxsw_sp_fib6_entry_offload_set(fib_entry);
  3362. break;
  3363. }
  3364. }
  3365. static void
  3366. mlxsw_sp_fib_entry_offload_unset(struct mlxsw_sp_fib_entry *fib_entry)
  3367. {
  3368. switch (fib_entry->fib_node->fib->proto) {
  3369. case MLXSW_SP_L3_PROTO_IPV4:
  3370. mlxsw_sp_fib4_entry_offload_unset(fib_entry);
  3371. break;
  3372. case MLXSW_SP_L3_PROTO_IPV6:
  3373. mlxsw_sp_fib6_entry_offload_unset(fib_entry);
  3374. break;
  3375. }
  3376. }
  3377. static void
  3378. mlxsw_sp_fib_entry_offload_refresh(struct mlxsw_sp_fib_entry *fib_entry,
  3379. enum mlxsw_reg_ralue_op op, int err)
  3380. {
  3381. switch (op) {
  3382. case MLXSW_REG_RALUE_OP_WRITE_DELETE:
  3383. return mlxsw_sp_fib_entry_offload_unset(fib_entry);
  3384. case MLXSW_REG_RALUE_OP_WRITE_WRITE:
  3385. if (err)
  3386. return;
  3387. if (mlxsw_sp_fib_entry_should_offload(fib_entry))
  3388. mlxsw_sp_fib_entry_offload_set(fib_entry);
  3389. else
  3390. mlxsw_sp_fib_entry_offload_unset(fib_entry);
  3391. return;
  3392. default:
  3393. return;
  3394. }
  3395. }
  3396. static void
  3397. mlxsw_sp_fib_entry_ralue_pack(char *ralue_pl,
  3398. const struct mlxsw_sp_fib_entry *fib_entry,
  3399. enum mlxsw_reg_ralue_op op)
  3400. {
  3401. struct mlxsw_sp_fib *fib = fib_entry->fib_node->fib;
  3402. enum mlxsw_reg_ralxx_protocol proto;
  3403. u32 *p_dip;
  3404. proto = (enum mlxsw_reg_ralxx_protocol) fib->proto;
  3405. switch (fib->proto) {
  3406. case MLXSW_SP_L3_PROTO_IPV4:
  3407. p_dip = (u32 *) fib_entry->fib_node->key.addr;
  3408. mlxsw_reg_ralue_pack4(ralue_pl, proto, op, fib->vr->id,
  3409. fib_entry->fib_node->key.prefix_len,
  3410. *p_dip);
  3411. break;
  3412. case MLXSW_SP_L3_PROTO_IPV6:
  3413. mlxsw_reg_ralue_pack6(ralue_pl, proto, op, fib->vr->id,
  3414. fib_entry->fib_node->key.prefix_len,
  3415. fib_entry->fib_node->key.addr);
  3416. break;
  3417. }
  3418. }
  3419. static int mlxsw_sp_fib_entry_op_remote(struct mlxsw_sp *mlxsw_sp,
  3420. struct mlxsw_sp_fib_entry *fib_entry,
  3421. enum mlxsw_reg_ralue_op op)
  3422. {
  3423. char ralue_pl[MLXSW_REG_RALUE_LEN];
  3424. enum mlxsw_reg_ralue_trap_action trap_action;
  3425. u16 trap_id = 0;
  3426. u32 adjacency_index = 0;
  3427. u16 ecmp_size = 0;
  3428. /* In case the nexthop group adjacency index is valid, use it
  3429. * with provided ECMP size. Otherwise, setup trap and pass
  3430. * traffic to kernel.
  3431. */
  3432. if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
  3433. trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
  3434. adjacency_index = fib_entry->nh_group->adj_index;
  3435. ecmp_size = fib_entry->nh_group->ecmp_size;
  3436. } else {
  3437. trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
  3438. trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
  3439. }
  3440. mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
  3441. mlxsw_reg_ralue_act_remote_pack(ralue_pl, trap_action, trap_id,
  3442. adjacency_index, ecmp_size);
  3443. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
  3444. }
  3445. static int mlxsw_sp_fib_entry_op_local(struct mlxsw_sp *mlxsw_sp,
  3446. struct mlxsw_sp_fib_entry *fib_entry,
  3447. enum mlxsw_reg_ralue_op op)
  3448. {
  3449. struct mlxsw_sp_rif *rif = fib_entry->nh_group->nh_rif;
  3450. enum mlxsw_reg_ralue_trap_action trap_action;
  3451. char ralue_pl[MLXSW_REG_RALUE_LEN];
  3452. u16 trap_id = 0;
  3453. u16 rif_index = 0;
  3454. if (mlxsw_sp_fib_entry_should_offload(fib_entry)) {
  3455. trap_action = MLXSW_REG_RALUE_TRAP_ACTION_NOP;
  3456. rif_index = rif->rif_index;
  3457. } else {
  3458. trap_action = MLXSW_REG_RALUE_TRAP_ACTION_TRAP;
  3459. trap_id = MLXSW_TRAP_ID_RTR_INGRESS0;
  3460. }
  3461. mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
  3462. mlxsw_reg_ralue_act_local_pack(ralue_pl, trap_action, trap_id,
  3463. rif_index);
  3464. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
  3465. }
  3466. static int mlxsw_sp_fib_entry_op_trap(struct mlxsw_sp *mlxsw_sp,
  3467. struct mlxsw_sp_fib_entry *fib_entry,
  3468. enum mlxsw_reg_ralue_op op)
  3469. {
  3470. char ralue_pl[MLXSW_REG_RALUE_LEN];
  3471. mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
  3472. mlxsw_reg_ralue_act_ip2me_pack(ralue_pl);
  3473. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue), ralue_pl);
  3474. }
  3475. static int
  3476. mlxsw_sp_fib_entry_op_ipip_decap(struct mlxsw_sp *mlxsw_sp,
  3477. struct mlxsw_sp_fib_entry *fib_entry,
  3478. enum mlxsw_reg_ralue_op op)
  3479. {
  3480. struct mlxsw_sp_ipip_entry *ipip_entry = fib_entry->decap.ipip_entry;
  3481. const struct mlxsw_sp_ipip_ops *ipip_ops;
  3482. if (WARN_ON(!ipip_entry))
  3483. return -EINVAL;
  3484. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
  3485. return ipip_ops->fib_entry_op(mlxsw_sp, ipip_entry, op,
  3486. fib_entry->decap.tunnel_index);
  3487. }
  3488. static int __mlxsw_sp_fib_entry_op(struct mlxsw_sp *mlxsw_sp,
  3489. struct mlxsw_sp_fib_entry *fib_entry,
  3490. enum mlxsw_reg_ralue_op op)
  3491. {
  3492. switch (fib_entry->type) {
  3493. case MLXSW_SP_FIB_ENTRY_TYPE_REMOTE:
  3494. return mlxsw_sp_fib_entry_op_remote(mlxsw_sp, fib_entry, op);
  3495. case MLXSW_SP_FIB_ENTRY_TYPE_LOCAL:
  3496. return mlxsw_sp_fib_entry_op_local(mlxsw_sp, fib_entry, op);
  3497. case MLXSW_SP_FIB_ENTRY_TYPE_TRAP:
  3498. return mlxsw_sp_fib_entry_op_trap(mlxsw_sp, fib_entry, op);
  3499. case MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP:
  3500. return mlxsw_sp_fib_entry_op_ipip_decap(mlxsw_sp,
  3501. fib_entry, op);
  3502. }
  3503. return -EINVAL;
  3504. }
  3505. static int mlxsw_sp_fib_entry_op(struct mlxsw_sp *mlxsw_sp,
  3506. struct mlxsw_sp_fib_entry *fib_entry,
  3507. enum mlxsw_reg_ralue_op op)
  3508. {
  3509. int err = __mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry, op);
  3510. mlxsw_sp_fib_entry_offload_refresh(fib_entry, op, err);
  3511. return err;
  3512. }
  3513. static int mlxsw_sp_fib_entry_update(struct mlxsw_sp *mlxsw_sp,
  3514. struct mlxsw_sp_fib_entry *fib_entry)
  3515. {
  3516. return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
  3517. MLXSW_REG_RALUE_OP_WRITE_WRITE);
  3518. }
  3519. static int mlxsw_sp_fib_entry_del(struct mlxsw_sp *mlxsw_sp,
  3520. struct mlxsw_sp_fib_entry *fib_entry)
  3521. {
  3522. return mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry,
  3523. MLXSW_REG_RALUE_OP_WRITE_DELETE);
  3524. }
  3525. static int
  3526. mlxsw_sp_fib4_entry_type_set(struct mlxsw_sp *mlxsw_sp,
  3527. const struct fib_entry_notifier_info *fen_info,
  3528. struct mlxsw_sp_fib_entry *fib_entry)
  3529. {
  3530. union mlxsw_sp_l3addr dip = { .addr4 = htonl(fen_info->dst) };
  3531. struct net_device *dev = fen_info->fi->fib_dev;
  3532. struct mlxsw_sp_ipip_entry *ipip_entry;
  3533. struct fib_info *fi = fen_info->fi;
  3534. switch (fen_info->type) {
  3535. case RTN_LOCAL:
  3536. ipip_entry = mlxsw_sp_ipip_entry_find_by_decap(mlxsw_sp, dev,
  3537. MLXSW_SP_L3_PROTO_IPV4, dip);
  3538. if (ipip_entry && ipip_entry->ol_dev->flags & IFF_UP) {
  3539. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP;
  3540. return mlxsw_sp_fib_entry_decap_init(mlxsw_sp,
  3541. fib_entry,
  3542. ipip_entry);
  3543. }
  3544. /* fall through */
  3545. case RTN_BROADCAST:
  3546. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
  3547. return 0;
  3548. case RTN_UNREACHABLE: /* fall through */
  3549. case RTN_BLACKHOLE: /* fall through */
  3550. case RTN_PROHIBIT:
  3551. /* Packets hitting these routes need to be trapped, but
  3552. * can do so with a lower priority than packets directed
  3553. * at the host, so use action type local instead of trap.
  3554. */
  3555. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
  3556. return 0;
  3557. case RTN_UNICAST:
  3558. if (mlxsw_sp_fi_is_gateway(mlxsw_sp, fi))
  3559. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
  3560. else
  3561. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
  3562. return 0;
  3563. default:
  3564. return -EINVAL;
  3565. }
  3566. }
  3567. static struct mlxsw_sp_fib4_entry *
  3568. mlxsw_sp_fib4_entry_create(struct mlxsw_sp *mlxsw_sp,
  3569. struct mlxsw_sp_fib_node *fib_node,
  3570. const struct fib_entry_notifier_info *fen_info)
  3571. {
  3572. struct mlxsw_sp_fib4_entry *fib4_entry;
  3573. struct mlxsw_sp_fib_entry *fib_entry;
  3574. int err;
  3575. fib4_entry = kzalloc(sizeof(*fib4_entry), GFP_KERNEL);
  3576. if (!fib4_entry)
  3577. return ERR_PTR(-ENOMEM);
  3578. fib_entry = &fib4_entry->common;
  3579. err = mlxsw_sp_fib4_entry_type_set(mlxsw_sp, fen_info, fib_entry);
  3580. if (err)
  3581. goto err_fib4_entry_type_set;
  3582. err = mlxsw_sp_nexthop4_group_get(mlxsw_sp, fib_entry, fen_info->fi);
  3583. if (err)
  3584. goto err_nexthop4_group_get;
  3585. fib4_entry->prio = fen_info->fi->fib_priority;
  3586. fib4_entry->tb_id = fen_info->tb_id;
  3587. fib4_entry->type = fen_info->type;
  3588. fib4_entry->tos = fen_info->tos;
  3589. fib_entry->fib_node = fib_node;
  3590. return fib4_entry;
  3591. err_nexthop4_group_get:
  3592. err_fib4_entry_type_set:
  3593. kfree(fib4_entry);
  3594. return ERR_PTR(err);
  3595. }
  3596. static void mlxsw_sp_fib4_entry_destroy(struct mlxsw_sp *mlxsw_sp,
  3597. struct mlxsw_sp_fib4_entry *fib4_entry)
  3598. {
  3599. mlxsw_sp_nexthop4_group_put(mlxsw_sp, &fib4_entry->common);
  3600. kfree(fib4_entry);
  3601. }
  3602. static struct mlxsw_sp_fib4_entry *
  3603. mlxsw_sp_fib4_entry_lookup(struct mlxsw_sp *mlxsw_sp,
  3604. const struct fib_entry_notifier_info *fen_info)
  3605. {
  3606. struct mlxsw_sp_fib4_entry *fib4_entry;
  3607. struct mlxsw_sp_fib_node *fib_node;
  3608. struct mlxsw_sp_fib *fib;
  3609. struct mlxsw_sp_vr *vr;
  3610. vr = mlxsw_sp_vr_find(mlxsw_sp, fen_info->tb_id);
  3611. if (!vr)
  3612. return NULL;
  3613. fib = mlxsw_sp_vr_fib(vr, MLXSW_SP_L3_PROTO_IPV4);
  3614. fib_node = mlxsw_sp_fib_node_lookup(fib, &fen_info->dst,
  3615. sizeof(fen_info->dst),
  3616. fen_info->dst_len);
  3617. if (!fib_node)
  3618. return NULL;
  3619. list_for_each_entry(fib4_entry, &fib_node->entry_list, common.list) {
  3620. if (fib4_entry->tb_id == fen_info->tb_id &&
  3621. fib4_entry->tos == fen_info->tos &&
  3622. fib4_entry->type == fen_info->type &&
  3623. mlxsw_sp_nexthop4_group_fi(fib4_entry->common.nh_group) ==
  3624. fen_info->fi) {
  3625. return fib4_entry;
  3626. }
  3627. }
  3628. return NULL;
  3629. }
  3630. static const struct rhashtable_params mlxsw_sp_fib_ht_params = {
  3631. .key_offset = offsetof(struct mlxsw_sp_fib_node, key),
  3632. .head_offset = offsetof(struct mlxsw_sp_fib_node, ht_node),
  3633. .key_len = sizeof(struct mlxsw_sp_fib_key),
  3634. .automatic_shrinking = true,
  3635. };
  3636. static int mlxsw_sp_fib_node_insert(struct mlxsw_sp_fib *fib,
  3637. struct mlxsw_sp_fib_node *fib_node)
  3638. {
  3639. return rhashtable_insert_fast(&fib->ht, &fib_node->ht_node,
  3640. mlxsw_sp_fib_ht_params);
  3641. }
  3642. static void mlxsw_sp_fib_node_remove(struct mlxsw_sp_fib *fib,
  3643. struct mlxsw_sp_fib_node *fib_node)
  3644. {
  3645. rhashtable_remove_fast(&fib->ht, &fib_node->ht_node,
  3646. mlxsw_sp_fib_ht_params);
  3647. }
  3648. static struct mlxsw_sp_fib_node *
  3649. mlxsw_sp_fib_node_lookup(struct mlxsw_sp_fib *fib, const void *addr,
  3650. size_t addr_len, unsigned char prefix_len)
  3651. {
  3652. struct mlxsw_sp_fib_key key;
  3653. memset(&key, 0, sizeof(key));
  3654. memcpy(key.addr, addr, addr_len);
  3655. key.prefix_len = prefix_len;
  3656. return rhashtable_lookup_fast(&fib->ht, &key, mlxsw_sp_fib_ht_params);
  3657. }
  3658. static struct mlxsw_sp_fib_node *
  3659. mlxsw_sp_fib_node_create(struct mlxsw_sp_fib *fib, const void *addr,
  3660. size_t addr_len, unsigned char prefix_len)
  3661. {
  3662. struct mlxsw_sp_fib_node *fib_node;
  3663. fib_node = kzalloc(sizeof(*fib_node), GFP_KERNEL);
  3664. if (!fib_node)
  3665. return NULL;
  3666. INIT_LIST_HEAD(&fib_node->entry_list);
  3667. list_add(&fib_node->list, &fib->node_list);
  3668. memcpy(fib_node->key.addr, addr, addr_len);
  3669. fib_node->key.prefix_len = prefix_len;
  3670. return fib_node;
  3671. }
  3672. static void mlxsw_sp_fib_node_destroy(struct mlxsw_sp_fib_node *fib_node)
  3673. {
  3674. list_del(&fib_node->list);
  3675. WARN_ON(!list_empty(&fib_node->entry_list));
  3676. kfree(fib_node);
  3677. }
  3678. static bool
  3679. mlxsw_sp_fib_node_entry_is_first(const struct mlxsw_sp_fib_node *fib_node,
  3680. const struct mlxsw_sp_fib_entry *fib_entry)
  3681. {
  3682. return list_first_entry(&fib_node->entry_list,
  3683. struct mlxsw_sp_fib_entry, list) == fib_entry;
  3684. }
  3685. static int mlxsw_sp_fib_lpm_tree_link(struct mlxsw_sp *mlxsw_sp,
  3686. struct mlxsw_sp_fib_node *fib_node)
  3687. {
  3688. struct mlxsw_sp_prefix_usage req_prefix_usage;
  3689. struct mlxsw_sp_fib *fib = fib_node->fib;
  3690. struct mlxsw_sp_lpm_tree *lpm_tree;
  3691. int err;
  3692. lpm_tree = mlxsw_sp->router->lpm.proto_trees[fib->proto];
  3693. if (lpm_tree->prefix_ref_count[fib_node->key.prefix_len] != 0)
  3694. goto out;
  3695. mlxsw_sp_prefix_usage_cpy(&req_prefix_usage, &lpm_tree->prefix_usage);
  3696. mlxsw_sp_prefix_usage_set(&req_prefix_usage, fib_node->key.prefix_len);
  3697. lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
  3698. fib->proto);
  3699. if (IS_ERR(lpm_tree))
  3700. return PTR_ERR(lpm_tree);
  3701. err = mlxsw_sp_vrs_lpm_tree_replace(mlxsw_sp, fib, lpm_tree);
  3702. if (err)
  3703. goto err_lpm_tree_replace;
  3704. out:
  3705. lpm_tree->prefix_ref_count[fib_node->key.prefix_len]++;
  3706. return 0;
  3707. err_lpm_tree_replace:
  3708. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  3709. return err;
  3710. }
  3711. static void mlxsw_sp_fib_lpm_tree_unlink(struct mlxsw_sp *mlxsw_sp,
  3712. struct mlxsw_sp_fib_node *fib_node)
  3713. {
  3714. struct mlxsw_sp_lpm_tree *lpm_tree = fib_node->fib->lpm_tree;
  3715. struct mlxsw_sp_prefix_usage req_prefix_usage;
  3716. struct mlxsw_sp_fib *fib = fib_node->fib;
  3717. int err;
  3718. if (--lpm_tree->prefix_ref_count[fib_node->key.prefix_len] != 0)
  3719. return;
  3720. /* Try to construct a new LPM tree from the current prefix usage
  3721. * minus the unused one. If we fail, continue using the old one.
  3722. */
  3723. mlxsw_sp_prefix_usage_cpy(&req_prefix_usage, &lpm_tree->prefix_usage);
  3724. mlxsw_sp_prefix_usage_clear(&req_prefix_usage,
  3725. fib_node->key.prefix_len);
  3726. lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, &req_prefix_usage,
  3727. fib->proto);
  3728. if (IS_ERR(lpm_tree))
  3729. return;
  3730. err = mlxsw_sp_vrs_lpm_tree_replace(mlxsw_sp, fib, lpm_tree);
  3731. if (err)
  3732. goto err_lpm_tree_replace;
  3733. return;
  3734. err_lpm_tree_replace:
  3735. mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
  3736. }
  3737. static int mlxsw_sp_fib_node_init(struct mlxsw_sp *mlxsw_sp,
  3738. struct mlxsw_sp_fib_node *fib_node,
  3739. struct mlxsw_sp_fib *fib)
  3740. {
  3741. int err;
  3742. err = mlxsw_sp_fib_node_insert(fib, fib_node);
  3743. if (err)
  3744. return err;
  3745. fib_node->fib = fib;
  3746. err = mlxsw_sp_fib_lpm_tree_link(mlxsw_sp, fib_node);
  3747. if (err)
  3748. goto err_fib_lpm_tree_link;
  3749. return 0;
  3750. err_fib_lpm_tree_link:
  3751. fib_node->fib = NULL;
  3752. mlxsw_sp_fib_node_remove(fib, fib_node);
  3753. return err;
  3754. }
  3755. static void mlxsw_sp_fib_node_fini(struct mlxsw_sp *mlxsw_sp,
  3756. struct mlxsw_sp_fib_node *fib_node)
  3757. {
  3758. struct mlxsw_sp_fib *fib = fib_node->fib;
  3759. mlxsw_sp_fib_lpm_tree_unlink(mlxsw_sp, fib_node);
  3760. fib_node->fib = NULL;
  3761. mlxsw_sp_fib_node_remove(fib, fib_node);
  3762. }
  3763. static struct mlxsw_sp_fib_node *
  3764. mlxsw_sp_fib_node_get(struct mlxsw_sp *mlxsw_sp, u32 tb_id, const void *addr,
  3765. size_t addr_len, unsigned char prefix_len,
  3766. enum mlxsw_sp_l3proto proto)
  3767. {
  3768. struct mlxsw_sp_fib_node *fib_node;
  3769. struct mlxsw_sp_fib *fib;
  3770. struct mlxsw_sp_vr *vr;
  3771. int err;
  3772. vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id, NULL);
  3773. if (IS_ERR(vr))
  3774. return ERR_CAST(vr);
  3775. fib = mlxsw_sp_vr_fib(vr, proto);
  3776. fib_node = mlxsw_sp_fib_node_lookup(fib, addr, addr_len, prefix_len);
  3777. if (fib_node)
  3778. return fib_node;
  3779. fib_node = mlxsw_sp_fib_node_create(fib, addr, addr_len, prefix_len);
  3780. if (!fib_node) {
  3781. err = -ENOMEM;
  3782. goto err_fib_node_create;
  3783. }
  3784. err = mlxsw_sp_fib_node_init(mlxsw_sp, fib_node, fib);
  3785. if (err)
  3786. goto err_fib_node_init;
  3787. return fib_node;
  3788. err_fib_node_init:
  3789. mlxsw_sp_fib_node_destroy(fib_node);
  3790. err_fib_node_create:
  3791. mlxsw_sp_vr_put(mlxsw_sp, vr);
  3792. return ERR_PTR(err);
  3793. }
  3794. static void mlxsw_sp_fib_node_put(struct mlxsw_sp *mlxsw_sp,
  3795. struct mlxsw_sp_fib_node *fib_node)
  3796. {
  3797. struct mlxsw_sp_vr *vr = fib_node->fib->vr;
  3798. if (!list_empty(&fib_node->entry_list))
  3799. return;
  3800. mlxsw_sp_fib_node_fini(mlxsw_sp, fib_node);
  3801. mlxsw_sp_fib_node_destroy(fib_node);
  3802. mlxsw_sp_vr_put(mlxsw_sp, vr);
  3803. }
  3804. static struct mlxsw_sp_fib4_entry *
  3805. mlxsw_sp_fib4_node_entry_find(const struct mlxsw_sp_fib_node *fib_node,
  3806. const struct mlxsw_sp_fib4_entry *new4_entry)
  3807. {
  3808. struct mlxsw_sp_fib4_entry *fib4_entry;
  3809. list_for_each_entry(fib4_entry, &fib_node->entry_list, common.list) {
  3810. if (fib4_entry->tb_id > new4_entry->tb_id)
  3811. continue;
  3812. if (fib4_entry->tb_id != new4_entry->tb_id)
  3813. break;
  3814. if (fib4_entry->tos > new4_entry->tos)
  3815. continue;
  3816. if (fib4_entry->prio >= new4_entry->prio ||
  3817. fib4_entry->tos < new4_entry->tos)
  3818. return fib4_entry;
  3819. }
  3820. return NULL;
  3821. }
  3822. static int
  3823. mlxsw_sp_fib4_node_list_append(struct mlxsw_sp_fib4_entry *fib4_entry,
  3824. struct mlxsw_sp_fib4_entry *new4_entry)
  3825. {
  3826. struct mlxsw_sp_fib_node *fib_node;
  3827. if (WARN_ON(!fib4_entry))
  3828. return -EINVAL;
  3829. fib_node = fib4_entry->common.fib_node;
  3830. list_for_each_entry_from(fib4_entry, &fib_node->entry_list,
  3831. common.list) {
  3832. if (fib4_entry->tb_id != new4_entry->tb_id ||
  3833. fib4_entry->tos != new4_entry->tos ||
  3834. fib4_entry->prio != new4_entry->prio)
  3835. break;
  3836. }
  3837. list_add_tail(&new4_entry->common.list, &fib4_entry->common.list);
  3838. return 0;
  3839. }
  3840. static int
  3841. mlxsw_sp_fib4_node_list_insert(struct mlxsw_sp_fib4_entry *new4_entry,
  3842. bool replace, bool append)
  3843. {
  3844. struct mlxsw_sp_fib_node *fib_node = new4_entry->common.fib_node;
  3845. struct mlxsw_sp_fib4_entry *fib4_entry;
  3846. fib4_entry = mlxsw_sp_fib4_node_entry_find(fib_node, new4_entry);
  3847. if (append)
  3848. return mlxsw_sp_fib4_node_list_append(fib4_entry, new4_entry);
  3849. if (replace && WARN_ON(!fib4_entry))
  3850. return -EINVAL;
  3851. /* Insert new entry before replaced one, so that we can later
  3852. * remove the second.
  3853. */
  3854. if (fib4_entry) {
  3855. list_add_tail(&new4_entry->common.list,
  3856. &fib4_entry->common.list);
  3857. } else {
  3858. struct mlxsw_sp_fib4_entry *last;
  3859. list_for_each_entry(last, &fib_node->entry_list, common.list) {
  3860. if (new4_entry->tb_id > last->tb_id)
  3861. break;
  3862. fib4_entry = last;
  3863. }
  3864. if (fib4_entry)
  3865. list_add(&new4_entry->common.list,
  3866. &fib4_entry->common.list);
  3867. else
  3868. list_add(&new4_entry->common.list,
  3869. &fib_node->entry_list);
  3870. }
  3871. return 0;
  3872. }
  3873. static void
  3874. mlxsw_sp_fib4_node_list_remove(struct mlxsw_sp_fib4_entry *fib4_entry)
  3875. {
  3876. list_del(&fib4_entry->common.list);
  3877. }
  3878. static int mlxsw_sp_fib_node_entry_add(struct mlxsw_sp *mlxsw_sp,
  3879. struct mlxsw_sp_fib_entry *fib_entry)
  3880. {
  3881. struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
  3882. if (!mlxsw_sp_fib_node_entry_is_first(fib_node, fib_entry))
  3883. return 0;
  3884. /* To prevent packet loss, overwrite the previously offloaded
  3885. * entry.
  3886. */
  3887. if (!list_is_singular(&fib_node->entry_list)) {
  3888. enum mlxsw_reg_ralue_op op = MLXSW_REG_RALUE_OP_WRITE_DELETE;
  3889. struct mlxsw_sp_fib_entry *n = list_next_entry(fib_entry, list);
  3890. mlxsw_sp_fib_entry_offload_refresh(n, op, 0);
  3891. }
  3892. return mlxsw_sp_fib_entry_update(mlxsw_sp, fib_entry);
  3893. }
  3894. static void mlxsw_sp_fib_node_entry_del(struct mlxsw_sp *mlxsw_sp,
  3895. struct mlxsw_sp_fib_entry *fib_entry)
  3896. {
  3897. struct mlxsw_sp_fib_node *fib_node = fib_entry->fib_node;
  3898. if (!mlxsw_sp_fib_node_entry_is_first(fib_node, fib_entry))
  3899. return;
  3900. /* Promote the next entry by overwriting the deleted entry */
  3901. if (!list_is_singular(&fib_node->entry_list)) {
  3902. struct mlxsw_sp_fib_entry *n = list_next_entry(fib_entry, list);
  3903. enum mlxsw_reg_ralue_op op = MLXSW_REG_RALUE_OP_WRITE_DELETE;
  3904. mlxsw_sp_fib_entry_update(mlxsw_sp, n);
  3905. mlxsw_sp_fib_entry_offload_refresh(fib_entry, op, 0);
  3906. return;
  3907. }
  3908. mlxsw_sp_fib_entry_del(mlxsw_sp, fib_entry);
  3909. }
  3910. static int mlxsw_sp_fib4_node_entry_link(struct mlxsw_sp *mlxsw_sp,
  3911. struct mlxsw_sp_fib4_entry *fib4_entry,
  3912. bool replace, bool append)
  3913. {
  3914. int err;
  3915. err = mlxsw_sp_fib4_node_list_insert(fib4_entry, replace, append);
  3916. if (err)
  3917. return err;
  3918. err = mlxsw_sp_fib_node_entry_add(mlxsw_sp, &fib4_entry->common);
  3919. if (err)
  3920. goto err_fib_node_entry_add;
  3921. return 0;
  3922. err_fib_node_entry_add:
  3923. mlxsw_sp_fib4_node_list_remove(fib4_entry);
  3924. return err;
  3925. }
  3926. static void
  3927. mlxsw_sp_fib4_node_entry_unlink(struct mlxsw_sp *mlxsw_sp,
  3928. struct mlxsw_sp_fib4_entry *fib4_entry)
  3929. {
  3930. mlxsw_sp_fib_node_entry_del(mlxsw_sp, &fib4_entry->common);
  3931. mlxsw_sp_fib4_node_list_remove(fib4_entry);
  3932. if (fib4_entry->common.type == MLXSW_SP_FIB_ENTRY_TYPE_IPIP_DECAP)
  3933. mlxsw_sp_fib_entry_decap_fini(mlxsw_sp, &fib4_entry->common);
  3934. }
  3935. static void mlxsw_sp_fib4_entry_replace(struct mlxsw_sp *mlxsw_sp,
  3936. struct mlxsw_sp_fib4_entry *fib4_entry,
  3937. bool replace)
  3938. {
  3939. struct mlxsw_sp_fib_node *fib_node = fib4_entry->common.fib_node;
  3940. struct mlxsw_sp_fib4_entry *replaced;
  3941. if (!replace)
  3942. return;
  3943. /* We inserted the new entry before replaced one */
  3944. replaced = list_next_entry(fib4_entry, common.list);
  3945. mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, replaced);
  3946. mlxsw_sp_fib4_entry_destroy(mlxsw_sp, replaced);
  3947. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  3948. }
  3949. static int
  3950. mlxsw_sp_router_fib4_add(struct mlxsw_sp *mlxsw_sp,
  3951. const struct fib_entry_notifier_info *fen_info,
  3952. bool replace, bool append)
  3953. {
  3954. struct mlxsw_sp_fib4_entry *fib4_entry;
  3955. struct mlxsw_sp_fib_node *fib_node;
  3956. int err;
  3957. if (mlxsw_sp->router->aborted)
  3958. return 0;
  3959. fib_node = mlxsw_sp_fib_node_get(mlxsw_sp, fen_info->tb_id,
  3960. &fen_info->dst, sizeof(fen_info->dst),
  3961. fen_info->dst_len,
  3962. MLXSW_SP_L3_PROTO_IPV4);
  3963. if (IS_ERR(fib_node)) {
  3964. dev_warn(mlxsw_sp->bus_info->dev, "Failed to get FIB node\n");
  3965. return PTR_ERR(fib_node);
  3966. }
  3967. fib4_entry = mlxsw_sp_fib4_entry_create(mlxsw_sp, fib_node, fen_info);
  3968. if (IS_ERR(fib4_entry)) {
  3969. dev_warn(mlxsw_sp->bus_info->dev, "Failed to create FIB entry\n");
  3970. err = PTR_ERR(fib4_entry);
  3971. goto err_fib4_entry_create;
  3972. }
  3973. err = mlxsw_sp_fib4_node_entry_link(mlxsw_sp, fib4_entry, replace,
  3974. append);
  3975. if (err) {
  3976. dev_warn(mlxsw_sp->bus_info->dev, "Failed to link FIB entry to node\n");
  3977. goto err_fib4_node_entry_link;
  3978. }
  3979. mlxsw_sp_fib4_entry_replace(mlxsw_sp, fib4_entry, replace);
  3980. return 0;
  3981. err_fib4_node_entry_link:
  3982. mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
  3983. err_fib4_entry_create:
  3984. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  3985. return err;
  3986. }
  3987. static void mlxsw_sp_router_fib4_del(struct mlxsw_sp *mlxsw_sp,
  3988. struct fib_entry_notifier_info *fen_info)
  3989. {
  3990. struct mlxsw_sp_fib4_entry *fib4_entry;
  3991. struct mlxsw_sp_fib_node *fib_node;
  3992. if (mlxsw_sp->router->aborted)
  3993. return;
  3994. fib4_entry = mlxsw_sp_fib4_entry_lookup(mlxsw_sp, fen_info);
  3995. if (WARN_ON(!fib4_entry))
  3996. return;
  3997. fib_node = fib4_entry->common.fib_node;
  3998. mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, fib4_entry);
  3999. mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
  4000. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4001. }
  4002. static bool mlxsw_sp_fib6_rt_should_ignore(const struct fib6_info *rt)
  4003. {
  4004. /* Packets with link-local destination IP arriving to the router
  4005. * are trapped to the CPU, so no need to program specific routes
  4006. * for them.
  4007. */
  4008. if (ipv6_addr_type(&rt->fib6_dst.addr) & IPV6_ADDR_LINKLOCAL)
  4009. return true;
  4010. /* Multicast routes aren't supported, so ignore them. Neighbour
  4011. * Discovery packets are specifically trapped.
  4012. */
  4013. if (ipv6_addr_type(&rt->fib6_dst.addr) & IPV6_ADDR_MULTICAST)
  4014. return true;
  4015. /* Cloned routes are irrelevant in the forwarding path. */
  4016. if (rt->fib6_flags & RTF_CACHE)
  4017. return true;
  4018. return false;
  4019. }
  4020. static struct mlxsw_sp_rt6 *mlxsw_sp_rt6_create(struct fib6_info *rt)
  4021. {
  4022. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4023. mlxsw_sp_rt6 = kzalloc(sizeof(*mlxsw_sp_rt6), GFP_KERNEL);
  4024. if (!mlxsw_sp_rt6)
  4025. return ERR_PTR(-ENOMEM);
  4026. /* In case of route replace, replaced route is deleted with
  4027. * no notification. Take reference to prevent accessing freed
  4028. * memory.
  4029. */
  4030. mlxsw_sp_rt6->rt = rt;
  4031. fib6_info_hold(rt);
  4032. return mlxsw_sp_rt6;
  4033. }
  4034. #if IS_ENABLED(CONFIG_IPV6)
  4035. static void mlxsw_sp_rt6_release(struct fib6_info *rt)
  4036. {
  4037. fib6_info_release(rt);
  4038. }
  4039. #else
  4040. static void mlxsw_sp_rt6_release(struct fib6_info *rt)
  4041. {
  4042. }
  4043. #endif
  4044. static void mlxsw_sp_rt6_destroy(struct mlxsw_sp_rt6 *mlxsw_sp_rt6)
  4045. {
  4046. mlxsw_sp_rt6_release(mlxsw_sp_rt6->rt);
  4047. kfree(mlxsw_sp_rt6);
  4048. }
  4049. static bool mlxsw_sp_fib6_rt_can_mp(const struct fib6_info *rt)
  4050. {
  4051. /* RTF_CACHE routes are ignored */
  4052. return (rt->fib6_flags & (RTF_GATEWAY | RTF_ADDRCONF)) == RTF_GATEWAY;
  4053. }
  4054. static struct fib6_info *
  4055. mlxsw_sp_fib6_entry_rt(const struct mlxsw_sp_fib6_entry *fib6_entry)
  4056. {
  4057. return list_first_entry(&fib6_entry->rt6_list, struct mlxsw_sp_rt6,
  4058. list)->rt;
  4059. }
  4060. static struct mlxsw_sp_fib6_entry *
  4061. mlxsw_sp_fib6_node_mp_entry_find(const struct mlxsw_sp_fib_node *fib_node,
  4062. const struct fib6_info *nrt, bool replace)
  4063. {
  4064. struct mlxsw_sp_fib6_entry *fib6_entry;
  4065. if (!mlxsw_sp_fib6_rt_can_mp(nrt) || replace)
  4066. return NULL;
  4067. list_for_each_entry(fib6_entry, &fib_node->entry_list, common.list) {
  4068. struct fib6_info *rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
  4069. /* RT6_TABLE_LOCAL and RT6_TABLE_MAIN share the same
  4070. * virtual router.
  4071. */
  4072. if (rt->fib6_table->tb6_id > nrt->fib6_table->tb6_id)
  4073. continue;
  4074. if (rt->fib6_table->tb6_id != nrt->fib6_table->tb6_id)
  4075. break;
  4076. if (rt->fib6_metric < nrt->fib6_metric)
  4077. continue;
  4078. if (rt->fib6_metric == nrt->fib6_metric &&
  4079. mlxsw_sp_fib6_rt_can_mp(rt))
  4080. return fib6_entry;
  4081. if (rt->fib6_metric > nrt->fib6_metric)
  4082. break;
  4083. }
  4084. return NULL;
  4085. }
  4086. static struct mlxsw_sp_rt6 *
  4087. mlxsw_sp_fib6_entry_rt_find(const struct mlxsw_sp_fib6_entry *fib6_entry,
  4088. const struct fib6_info *rt)
  4089. {
  4090. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4091. list_for_each_entry(mlxsw_sp_rt6, &fib6_entry->rt6_list, list) {
  4092. if (mlxsw_sp_rt6->rt == rt)
  4093. return mlxsw_sp_rt6;
  4094. }
  4095. return NULL;
  4096. }
  4097. static bool mlxsw_sp_nexthop6_ipip_type(const struct mlxsw_sp *mlxsw_sp,
  4098. const struct fib6_info *rt,
  4099. enum mlxsw_sp_ipip_type *ret)
  4100. {
  4101. return rt->fib6_nh.nh_dev &&
  4102. mlxsw_sp_netdev_ipip_type(mlxsw_sp, rt->fib6_nh.nh_dev, ret);
  4103. }
  4104. static int mlxsw_sp_nexthop6_type_init(struct mlxsw_sp *mlxsw_sp,
  4105. struct mlxsw_sp_nexthop_group *nh_grp,
  4106. struct mlxsw_sp_nexthop *nh,
  4107. const struct fib6_info *rt)
  4108. {
  4109. const struct mlxsw_sp_ipip_ops *ipip_ops;
  4110. struct mlxsw_sp_ipip_entry *ipip_entry;
  4111. struct net_device *dev = rt->fib6_nh.nh_dev;
  4112. struct mlxsw_sp_rif *rif;
  4113. int err;
  4114. ipip_entry = mlxsw_sp_ipip_entry_find_by_ol_dev(mlxsw_sp, dev);
  4115. if (ipip_entry) {
  4116. ipip_ops = mlxsw_sp->router->ipip_ops_arr[ipip_entry->ipipt];
  4117. if (ipip_ops->can_offload(mlxsw_sp, dev,
  4118. MLXSW_SP_L3_PROTO_IPV6)) {
  4119. nh->type = MLXSW_SP_NEXTHOP_TYPE_IPIP;
  4120. mlxsw_sp_nexthop_ipip_init(mlxsw_sp, nh, ipip_entry);
  4121. return 0;
  4122. }
  4123. }
  4124. nh->type = MLXSW_SP_NEXTHOP_TYPE_ETH;
  4125. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  4126. if (!rif)
  4127. return 0;
  4128. mlxsw_sp_nexthop_rif_init(nh, rif);
  4129. err = mlxsw_sp_nexthop_neigh_init(mlxsw_sp, nh);
  4130. if (err)
  4131. goto err_nexthop_neigh_init;
  4132. return 0;
  4133. err_nexthop_neigh_init:
  4134. mlxsw_sp_nexthop_rif_fini(nh);
  4135. return err;
  4136. }
  4137. static void mlxsw_sp_nexthop6_type_fini(struct mlxsw_sp *mlxsw_sp,
  4138. struct mlxsw_sp_nexthop *nh)
  4139. {
  4140. mlxsw_sp_nexthop_type_fini(mlxsw_sp, nh);
  4141. }
  4142. static int mlxsw_sp_nexthop6_init(struct mlxsw_sp *mlxsw_sp,
  4143. struct mlxsw_sp_nexthop_group *nh_grp,
  4144. struct mlxsw_sp_nexthop *nh,
  4145. const struct fib6_info *rt)
  4146. {
  4147. struct net_device *dev = rt->fib6_nh.nh_dev;
  4148. nh->nh_grp = nh_grp;
  4149. nh->nh_weight = rt->fib6_nh.nh_weight;
  4150. memcpy(&nh->gw_addr, &rt->fib6_nh.nh_gw, sizeof(nh->gw_addr));
  4151. mlxsw_sp_nexthop_counter_alloc(mlxsw_sp, nh);
  4152. list_add_tail(&nh->router_list_node, &mlxsw_sp->router->nexthop_list);
  4153. if (!dev)
  4154. return 0;
  4155. nh->ifindex = dev->ifindex;
  4156. return mlxsw_sp_nexthop6_type_init(mlxsw_sp, nh_grp, nh, rt);
  4157. }
  4158. static void mlxsw_sp_nexthop6_fini(struct mlxsw_sp *mlxsw_sp,
  4159. struct mlxsw_sp_nexthop *nh)
  4160. {
  4161. mlxsw_sp_nexthop6_type_fini(mlxsw_sp, nh);
  4162. list_del(&nh->router_list_node);
  4163. mlxsw_sp_nexthop_counter_free(mlxsw_sp, nh);
  4164. }
  4165. static bool mlxsw_sp_rt6_is_gateway(const struct mlxsw_sp *mlxsw_sp,
  4166. const struct fib6_info *rt)
  4167. {
  4168. return rt->fib6_flags & RTF_GATEWAY ||
  4169. mlxsw_sp_nexthop6_ipip_type(mlxsw_sp, rt, NULL);
  4170. }
  4171. static struct mlxsw_sp_nexthop_group *
  4172. mlxsw_sp_nexthop6_group_create(struct mlxsw_sp *mlxsw_sp,
  4173. struct mlxsw_sp_fib6_entry *fib6_entry)
  4174. {
  4175. struct mlxsw_sp_nexthop_group *nh_grp;
  4176. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4177. struct mlxsw_sp_nexthop *nh;
  4178. size_t alloc_size;
  4179. int i = 0;
  4180. int err;
  4181. alloc_size = sizeof(*nh_grp) +
  4182. fib6_entry->nrt6 * sizeof(struct mlxsw_sp_nexthop);
  4183. nh_grp = kzalloc(alloc_size, GFP_KERNEL);
  4184. if (!nh_grp)
  4185. return ERR_PTR(-ENOMEM);
  4186. INIT_LIST_HEAD(&nh_grp->fib_list);
  4187. #if IS_ENABLED(CONFIG_IPV6)
  4188. nh_grp->neigh_tbl = &nd_tbl;
  4189. #endif
  4190. mlxsw_sp_rt6 = list_first_entry(&fib6_entry->rt6_list,
  4191. struct mlxsw_sp_rt6, list);
  4192. nh_grp->gateway = mlxsw_sp_rt6_is_gateway(mlxsw_sp, mlxsw_sp_rt6->rt);
  4193. nh_grp->count = fib6_entry->nrt6;
  4194. for (i = 0; i < nh_grp->count; i++) {
  4195. struct fib6_info *rt = mlxsw_sp_rt6->rt;
  4196. nh = &nh_grp->nexthops[i];
  4197. err = mlxsw_sp_nexthop6_init(mlxsw_sp, nh_grp, nh, rt);
  4198. if (err)
  4199. goto err_nexthop6_init;
  4200. mlxsw_sp_rt6 = list_next_entry(mlxsw_sp_rt6, list);
  4201. }
  4202. err = mlxsw_sp_nexthop_group_insert(mlxsw_sp, nh_grp);
  4203. if (err)
  4204. goto err_nexthop_group_insert;
  4205. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
  4206. return nh_grp;
  4207. err_nexthop_group_insert:
  4208. err_nexthop6_init:
  4209. for (i--; i >= 0; i--) {
  4210. nh = &nh_grp->nexthops[i];
  4211. mlxsw_sp_nexthop6_fini(mlxsw_sp, nh);
  4212. }
  4213. kfree(nh_grp);
  4214. return ERR_PTR(err);
  4215. }
  4216. static void
  4217. mlxsw_sp_nexthop6_group_destroy(struct mlxsw_sp *mlxsw_sp,
  4218. struct mlxsw_sp_nexthop_group *nh_grp)
  4219. {
  4220. struct mlxsw_sp_nexthop *nh;
  4221. int i = nh_grp->count;
  4222. mlxsw_sp_nexthop_group_remove(mlxsw_sp, nh_grp);
  4223. for (i--; i >= 0; i--) {
  4224. nh = &nh_grp->nexthops[i];
  4225. mlxsw_sp_nexthop6_fini(mlxsw_sp, nh);
  4226. }
  4227. mlxsw_sp_nexthop_group_refresh(mlxsw_sp, nh_grp);
  4228. WARN_ON(nh_grp->adj_index_valid);
  4229. kfree(nh_grp);
  4230. }
  4231. static int mlxsw_sp_nexthop6_group_get(struct mlxsw_sp *mlxsw_sp,
  4232. struct mlxsw_sp_fib6_entry *fib6_entry)
  4233. {
  4234. struct mlxsw_sp_nexthop_group *nh_grp;
  4235. nh_grp = mlxsw_sp_nexthop6_group_lookup(mlxsw_sp, fib6_entry);
  4236. if (!nh_grp) {
  4237. nh_grp = mlxsw_sp_nexthop6_group_create(mlxsw_sp, fib6_entry);
  4238. if (IS_ERR(nh_grp))
  4239. return PTR_ERR(nh_grp);
  4240. }
  4241. list_add_tail(&fib6_entry->common.nexthop_group_node,
  4242. &nh_grp->fib_list);
  4243. fib6_entry->common.nh_group = nh_grp;
  4244. return 0;
  4245. }
  4246. static void mlxsw_sp_nexthop6_group_put(struct mlxsw_sp *mlxsw_sp,
  4247. struct mlxsw_sp_fib_entry *fib_entry)
  4248. {
  4249. struct mlxsw_sp_nexthop_group *nh_grp = fib_entry->nh_group;
  4250. list_del(&fib_entry->nexthop_group_node);
  4251. if (!list_empty(&nh_grp->fib_list))
  4252. return;
  4253. mlxsw_sp_nexthop6_group_destroy(mlxsw_sp, nh_grp);
  4254. }
  4255. static int
  4256. mlxsw_sp_nexthop6_group_update(struct mlxsw_sp *mlxsw_sp,
  4257. struct mlxsw_sp_fib6_entry *fib6_entry)
  4258. {
  4259. struct mlxsw_sp_nexthop_group *old_nh_grp = fib6_entry->common.nh_group;
  4260. int err;
  4261. fib6_entry->common.nh_group = NULL;
  4262. list_del(&fib6_entry->common.nexthop_group_node);
  4263. err = mlxsw_sp_nexthop6_group_get(mlxsw_sp, fib6_entry);
  4264. if (err)
  4265. goto err_nexthop6_group_get;
  4266. /* In case this entry is offloaded, then the adjacency index
  4267. * currently associated with it in the device's table is that
  4268. * of the old group. Start using the new one instead.
  4269. */
  4270. err = mlxsw_sp_fib_node_entry_add(mlxsw_sp, &fib6_entry->common);
  4271. if (err)
  4272. goto err_fib_node_entry_add;
  4273. if (list_empty(&old_nh_grp->fib_list))
  4274. mlxsw_sp_nexthop6_group_destroy(mlxsw_sp, old_nh_grp);
  4275. return 0;
  4276. err_fib_node_entry_add:
  4277. mlxsw_sp_nexthop6_group_put(mlxsw_sp, &fib6_entry->common);
  4278. err_nexthop6_group_get:
  4279. list_add_tail(&fib6_entry->common.nexthop_group_node,
  4280. &old_nh_grp->fib_list);
  4281. fib6_entry->common.nh_group = old_nh_grp;
  4282. return err;
  4283. }
  4284. static int
  4285. mlxsw_sp_fib6_entry_nexthop_add(struct mlxsw_sp *mlxsw_sp,
  4286. struct mlxsw_sp_fib6_entry *fib6_entry,
  4287. struct fib6_info *rt)
  4288. {
  4289. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4290. int err;
  4291. mlxsw_sp_rt6 = mlxsw_sp_rt6_create(rt);
  4292. if (IS_ERR(mlxsw_sp_rt6))
  4293. return PTR_ERR(mlxsw_sp_rt6);
  4294. list_add_tail(&mlxsw_sp_rt6->list, &fib6_entry->rt6_list);
  4295. fib6_entry->nrt6++;
  4296. err = mlxsw_sp_nexthop6_group_update(mlxsw_sp, fib6_entry);
  4297. if (err)
  4298. goto err_nexthop6_group_update;
  4299. return 0;
  4300. err_nexthop6_group_update:
  4301. fib6_entry->nrt6--;
  4302. list_del(&mlxsw_sp_rt6->list);
  4303. mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
  4304. return err;
  4305. }
  4306. static void
  4307. mlxsw_sp_fib6_entry_nexthop_del(struct mlxsw_sp *mlxsw_sp,
  4308. struct mlxsw_sp_fib6_entry *fib6_entry,
  4309. struct fib6_info *rt)
  4310. {
  4311. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4312. mlxsw_sp_rt6 = mlxsw_sp_fib6_entry_rt_find(fib6_entry, rt);
  4313. if (WARN_ON(!mlxsw_sp_rt6))
  4314. return;
  4315. fib6_entry->nrt6--;
  4316. list_del(&mlxsw_sp_rt6->list);
  4317. mlxsw_sp_nexthop6_group_update(mlxsw_sp, fib6_entry);
  4318. mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
  4319. }
  4320. static void mlxsw_sp_fib6_entry_type_set(struct mlxsw_sp *mlxsw_sp,
  4321. struct mlxsw_sp_fib_entry *fib_entry,
  4322. const struct fib6_info *rt)
  4323. {
  4324. /* Packets hitting RTF_REJECT routes need to be discarded by the
  4325. * stack. We can rely on their destination device not having a
  4326. * RIF (it's the loopback device) and can thus use action type
  4327. * local, which will cause them to be trapped with a lower
  4328. * priority than packets that need to be locally received.
  4329. */
  4330. if (rt->fib6_flags & (RTF_LOCAL | RTF_ANYCAST))
  4331. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_TRAP;
  4332. else if (rt->fib6_flags & RTF_REJECT)
  4333. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
  4334. else if (mlxsw_sp_rt6_is_gateway(mlxsw_sp, rt))
  4335. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_REMOTE;
  4336. else
  4337. fib_entry->type = MLXSW_SP_FIB_ENTRY_TYPE_LOCAL;
  4338. }
  4339. static void
  4340. mlxsw_sp_fib6_entry_rt_destroy_all(struct mlxsw_sp_fib6_entry *fib6_entry)
  4341. {
  4342. struct mlxsw_sp_rt6 *mlxsw_sp_rt6, *tmp;
  4343. list_for_each_entry_safe(mlxsw_sp_rt6, tmp, &fib6_entry->rt6_list,
  4344. list) {
  4345. fib6_entry->nrt6--;
  4346. list_del(&mlxsw_sp_rt6->list);
  4347. mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
  4348. }
  4349. }
  4350. static struct mlxsw_sp_fib6_entry *
  4351. mlxsw_sp_fib6_entry_create(struct mlxsw_sp *mlxsw_sp,
  4352. struct mlxsw_sp_fib_node *fib_node,
  4353. struct fib6_info *rt)
  4354. {
  4355. struct mlxsw_sp_fib6_entry *fib6_entry;
  4356. struct mlxsw_sp_fib_entry *fib_entry;
  4357. struct mlxsw_sp_rt6 *mlxsw_sp_rt6;
  4358. int err;
  4359. fib6_entry = kzalloc(sizeof(*fib6_entry), GFP_KERNEL);
  4360. if (!fib6_entry)
  4361. return ERR_PTR(-ENOMEM);
  4362. fib_entry = &fib6_entry->common;
  4363. mlxsw_sp_rt6 = mlxsw_sp_rt6_create(rt);
  4364. if (IS_ERR(mlxsw_sp_rt6)) {
  4365. err = PTR_ERR(mlxsw_sp_rt6);
  4366. goto err_rt6_create;
  4367. }
  4368. mlxsw_sp_fib6_entry_type_set(mlxsw_sp, fib_entry, mlxsw_sp_rt6->rt);
  4369. INIT_LIST_HEAD(&fib6_entry->rt6_list);
  4370. list_add_tail(&mlxsw_sp_rt6->list, &fib6_entry->rt6_list);
  4371. fib6_entry->nrt6 = 1;
  4372. err = mlxsw_sp_nexthop6_group_get(mlxsw_sp, fib6_entry);
  4373. if (err)
  4374. goto err_nexthop6_group_get;
  4375. fib_entry->fib_node = fib_node;
  4376. return fib6_entry;
  4377. err_nexthop6_group_get:
  4378. list_del(&mlxsw_sp_rt6->list);
  4379. mlxsw_sp_rt6_destroy(mlxsw_sp_rt6);
  4380. err_rt6_create:
  4381. kfree(fib6_entry);
  4382. return ERR_PTR(err);
  4383. }
  4384. static void mlxsw_sp_fib6_entry_destroy(struct mlxsw_sp *mlxsw_sp,
  4385. struct mlxsw_sp_fib6_entry *fib6_entry)
  4386. {
  4387. mlxsw_sp_nexthop6_group_put(mlxsw_sp, &fib6_entry->common);
  4388. mlxsw_sp_fib6_entry_rt_destroy_all(fib6_entry);
  4389. WARN_ON(fib6_entry->nrt6);
  4390. kfree(fib6_entry);
  4391. }
  4392. static struct mlxsw_sp_fib6_entry *
  4393. mlxsw_sp_fib6_node_entry_find(const struct mlxsw_sp_fib_node *fib_node,
  4394. const struct fib6_info *nrt, bool replace)
  4395. {
  4396. struct mlxsw_sp_fib6_entry *fib6_entry, *fallback = NULL;
  4397. list_for_each_entry(fib6_entry, &fib_node->entry_list, common.list) {
  4398. struct fib6_info *rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
  4399. if (rt->fib6_table->tb6_id > nrt->fib6_table->tb6_id)
  4400. continue;
  4401. if (rt->fib6_table->tb6_id != nrt->fib6_table->tb6_id)
  4402. break;
  4403. if (replace && rt->fib6_metric == nrt->fib6_metric) {
  4404. if (mlxsw_sp_fib6_rt_can_mp(rt) ==
  4405. mlxsw_sp_fib6_rt_can_mp(nrt))
  4406. return fib6_entry;
  4407. if (mlxsw_sp_fib6_rt_can_mp(nrt))
  4408. fallback = fallback ?: fib6_entry;
  4409. }
  4410. if (rt->fib6_metric > nrt->fib6_metric)
  4411. return fallback ?: fib6_entry;
  4412. }
  4413. return fallback;
  4414. }
  4415. static int
  4416. mlxsw_sp_fib6_node_list_insert(struct mlxsw_sp_fib6_entry *new6_entry,
  4417. bool replace)
  4418. {
  4419. struct mlxsw_sp_fib_node *fib_node = new6_entry->common.fib_node;
  4420. struct fib6_info *nrt = mlxsw_sp_fib6_entry_rt(new6_entry);
  4421. struct mlxsw_sp_fib6_entry *fib6_entry;
  4422. fib6_entry = mlxsw_sp_fib6_node_entry_find(fib_node, nrt, replace);
  4423. if (replace && WARN_ON(!fib6_entry))
  4424. return -EINVAL;
  4425. if (fib6_entry) {
  4426. list_add_tail(&new6_entry->common.list,
  4427. &fib6_entry->common.list);
  4428. } else {
  4429. struct mlxsw_sp_fib6_entry *last;
  4430. list_for_each_entry(last, &fib_node->entry_list, common.list) {
  4431. struct fib6_info *rt = mlxsw_sp_fib6_entry_rt(last);
  4432. if (nrt->fib6_table->tb6_id > rt->fib6_table->tb6_id)
  4433. break;
  4434. fib6_entry = last;
  4435. }
  4436. if (fib6_entry)
  4437. list_add(&new6_entry->common.list,
  4438. &fib6_entry->common.list);
  4439. else
  4440. list_add(&new6_entry->common.list,
  4441. &fib_node->entry_list);
  4442. }
  4443. return 0;
  4444. }
  4445. static void
  4446. mlxsw_sp_fib6_node_list_remove(struct mlxsw_sp_fib6_entry *fib6_entry)
  4447. {
  4448. list_del(&fib6_entry->common.list);
  4449. }
  4450. static int mlxsw_sp_fib6_node_entry_link(struct mlxsw_sp *mlxsw_sp,
  4451. struct mlxsw_sp_fib6_entry *fib6_entry,
  4452. bool replace)
  4453. {
  4454. int err;
  4455. err = mlxsw_sp_fib6_node_list_insert(fib6_entry, replace);
  4456. if (err)
  4457. return err;
  4458. err = mlxsw_sp_fib_node_entry_add(mlxsw_sp, &fib6_entry->common);
  4459. if (err)
  4460. goto err_fib_node_entry_add;
  4461. return 0;
  4462. err_fib_node_entry_add:
  4463. mlxsw_sp_fib6_node_list_remove(fib6_entry);
  4464. return err;
  4465. }
  4466. static void
  4467. mlxsw_sp_fib6_node_entry_unlink(struct mlxsw_sp *mlxsw_sp,
  4468. struct mlxsw_sp_fib6_entry *fib6_entry)
  4469. {
  4470. mlxsw_sp_fib_node_entry_del(mlxsw_sp, &fib6_entry->common);
  4471. mlxsw_sp_fib6_node_list_remove(fib6_entry);
  4472. }
  4473. static struct mlxsw_sp_fib6_entry *
  4474. mlxsw_sp_fib6_entry_lookup(struct mlxsw_sp *mlxsw_sp,
  4475. const struct fib6_info *rt)
  4476. {
  4477. struct mlxsw_sp_fib6_entry *fib6_entry;
  4478. struct mlxsw_sp_fib_node *fib_node;
  4479. struct mlxsw_sp_fib *fib;
  4480. struct mlxsw_sp_vr *vr;
  4481. vr = mlxsw_sp_vr_find(mlxsw_sp, rt->fib6_table->tb6_id);
  4482. if (!vr)
  4483. return NULL;
  4484. fib = mlxsw_sp_vr_fib(vr, MLXSW_SP_L3_PROTO_IPV6);
  4485. fib_node = mlxsw_sp_fib_node_lookup(fib, &rt->fib6_dst.addr,
  4486. sizeof(rt->fib6_dst.addr),
  4487. rt->fib6_dst.plen);
  4488. if (!fib_node)
  4489. return NULL;
  4490. list_for_each_entry(fib6_entry, &fib_node->entry_list, common.list) {
  4491. struct fib6_info *iter_rt = mlxsw_sp_fib6_entry_rt(fib6_entry);
  4492. if (rt->fib6_table->tb6_id == iter_rt->fib6_table->tb6_id &&
  4493. rt->fib6_metric == iter_rt->fib6_metric &&
  4494. mlxsw_sp_fib6_entry_rt_find(fib6_entry, rt))
  4495. return fib6_entry;
  4496. }
  4497. return NULL;
  4498. }
  4499. static void mlxsw_sp_fib6_entry_replace(struct mlxsw_sp *mlxsw_sp,
  4500. struct mlxsw_sp_fib6_entry *fib6_entry,
  4501. bool replace)
  4502. {
  4503. struct mlxsw_sp_fib_node *fib_node = fib6_entry->common.fib_node;
  4504. struct mlxsw_sp_fib6_entry *replaced;
  4505. if (!replace)
  4506. return;
  4507. replaced = list_next_entry(fib6_entry, common.list);
  4508. mlxsw_sp_fib6_node_entry_unlink(mlxsw_sp, replaced);
  4509. mlxsw_sp_fib6_entry_destroy(mlxsw_sp, replaced);
  4510. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4511. }
  4512. static int mlxsw_sp_router_fib6_add(struct mlxsw_sp *mlxsw_sp,
  4513. struct fib6_info *rt, bool replace)
  4514. {
  4515. struct mlxsw_sp_fib6_entry *fib6_entry;
  4516. struct mlxsw_sp_fib_node *fib_node;
  4517. int err;
  4518. if (mlxsw_sp->router->aborted)
  4519. return 0;
  4520. if (rt->fib6_src.plen)
  4521. return -EINVAL;
  4522. if (mlxsw_sp_fib6_rt_should_ignore(rt))
  4523. return 0;
  4524. fib_node = mlxsw_sp_fib_node_get(mlxsw_sp, rt->fib6_table->tb6_id,
  4525. &rt->fib6_dst.addr,
  4526. sizeof(rt->fib6_dst.addr),
  4527. rt->fib6_dst.plen,
  4528. MLXSW_SP_L3_PROTO_IPV6);
  4529. if (IS_ERR(fib_node))
  4530. return PTR_ERR(fib_node);
  4531. /* Before creating a new entry, try to append route to an existing
  4532. * multipath entry.
  4533. */
  4534. fib6_entry = mlxsw_sp_fib6_node_mp_entry_find(fib_node, rt, replace);
  4535. if (fib6_entry) {
  4536. err = mlxsw_sp_fib6_entry_nexthop_add(mlxsw_sp, fib6_entry, rt);
  4537. if (err)
  4538. goto err_fib6_entry_nexthop_add;
  4539. return 0;
  4540. }
  4541. fib6_entry = mlxsw_sp_fib6_entry_create(mlxsw_sp, fib_node, rt);
  4542. if (IS_ERR(fib6_entry)) {
  4543. err = PTR_ERR(fib6_entry);
  4544. goto err_fib6_entry_create;
  4545. }
  4546. err = mlxsw_sp_fib6_node_entry_link(mlxsw_sp, fib6_entry, replace);
  4547. if (err)
  4548. goto err_fib6_node_entry_link;
  4549. mlxsw_sp_fib6_entry_replace(mlxsw_sp, fib6_entry, replace);
  4550. return 0;
  4551. err_fib6_node_entry_link:
  4552. mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
  4553. err_fib6_entry_create:
  4554. err_fib6_entry_nexthop_add:
  4555. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4556. return err;
  4557. }
  4558. static void mlxsw_sp_router_fib6_del(struct mlxsw_sp *mlxsw_sp,
  4559. struct fib6_info *rt)
  4560. {
  4561. struct mlxsw_sp_fib6_entry *fib6_entry;
  4562. struct mlxsw_sp_fib_node *fib_node;
  4563. if (mlxsw_sp->router->aborted)
  4564. return;
  4565. if (mlxsw_sp_fib6_rt_should_ignore(rt))
  4566. return;
  4567. fib6_entry = mlxsw_sp_fib6_entry_lookup(mlxsw_sp, rt);
  4568. if (WARN_ON(!fib6_entry))
  4569. return;
  4570. /* If route is part of a multipath entry, but not the last one
  4571. * removed, then only reduce its nexthop group.
  4572. */
  4573. if (!list_is_singular(&fib6_entry->rt6_list)) {
  4574. mlxsw_sp_fib6_entry_nexthop_del(mlxsw_sp, fib6_entry, rt);
  4575. return;
  4576. }
  4577. fib_node = fib6_entry->common.fib_node;
  4578. mlxsw_sp_fib6_node_entry_unlink(mlxsw_sp, fib6_entry);
  4579. mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
  4580. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4581. }
  4582. static int __mlxsw_sp_router_set_abort_trap(struct mlxsw_sp *mlxsw_sp,
  4583. enum mlxsw_reg_ralxx_protocol proto,
  4584. u8 tree_id)
  4585. {
  4586. char ralta_pl[MLXSW_REG_RALTA_LEN];
  4587. char ralst_pl[MLXSW_REG_RALST_LEN];
  4588. int i, err;
  4589. mlxsw_reg_ralta_pack(ralta_pl, true, proto, tree_id);
  4590. err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralta), ralta_pl);
  4591. if (err)
  4592. return err;
  4593. mlxsw_reg_ralst_pack(ralst_pl, 0xff, tree_id);
  4594. err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralst), ralst_pl);
  4595. if (err)
  4596. return err;
  4597. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
  4598. struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[i];
  4599. char raltb_pl[MLXSW_REG_RALTB_LEN];
  4600. char ralue_pl[MLXSW_REG_RALUE_LEN];
  4601. mlxsw_reg_raltb_pack(raltb_pl, vr->id, proto, tree_id);
  4602. err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(raltb),
  4603. raltb_pl);
  4604. if (err)
  4605. return err;
  4606. mlxsw_reg_ralue_pack(ralue_pl, proto,
  4607. MLXSW_REG_RALUE_OP_WRITE_WRITE, vr->id, 0);
  4608. mlxsw_reg_ralue_act_ip2me_pack(ralue_pl);
  4609. err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ralue),
  4610. ralue_pl);
  4611. if (err)
  4612. return err;
  4613. }
  4614. return 0;
  4615. }
  4616. static struct mlxsw_sp_mr_table *
  4617. mlxsw_sp_router_fibmr_family_to_table(struct mlxsw_sp_vr *vr, int family)
  4618. {
  4619. if (family == RTNL_FAMILY_IPMR)
  4620. return vr->mr_table[MLXSW_SP_L3_PROTO_IPV4];
  4621. else
  4622. return vr->mr_table[MLXSW_SP_L3_PROTO_IPV6];
  4623. }
  4624. static int mlxsw_sp_router_fibmr_add(struct mlxsw_sp *mlxsw_sp,
  4625. struct mfc_entry_notifier_info *men_info,
  4626. bool replace)
  4627. {
  4628. struct mlxsw_sp_mr_table *mrt;
  4629. struct mlxsw_sp_vr *vr;
  4630. if (mlxsw_sp->router->aborted)
  4631. return 0;
  4632. vr = mlxsw_sp_vr_get(mlxsw_sp, men_info->tb_id, NULL);
  4633. if (IS_ERR(vr))
  4634. return PTR_ERR(vr);
  4635. mrt = mlxsw_sp_router_fibmr_family_to_table(vr, men_info->info.family);
  4636. return mlxsw_sp_mr_route_add(mrt, men_info->mfc, replace);
  4637. }
  4638. static void mlxsw_sp_router_fibmr_del(struct mlxsw_sp *mlxsw_sp,
  4639. struct mfc_entry_notifier_info *men_info)
  4640. {
  4641. struct mlxsw_sp_mr_table *mrt;
  4642. struct mlxsw_sp_vr *vr;
  4643. if (mlxsw_sp->router->aborted)
  4644. return;
  4645. vr = mlxsw_sp_vr_find(mlxsw_sp, men_info->tb_id);
  4646. if (WARN_ON(!vr))
  4647. return;
  4648. mrt = mlxsw_sp_router_fibmr_family_to_table(vr, men_info->info.family);
  4649. mlxsw_sp_mr_route_del(mrt, men_info->mfc);
  4650. mlxsw_sp_vr_put(mlxsw_sp, vr);
  4651. }
  4652. static int
  4653. mlxsw_sp_router_fibmr_vif_add(struct mlxsw_sp *mlxsw_sp,
  4654. struct vif_entry_notifier_info *ven_info)
  4655. {
  4656. struct mlxsw_sp_mr_table *mrt;
  4657. struct mlxsw_sp_rif *rif;
  4658. struct mlxsw_sp_vr *vr;
  4659. if (mlxsw_sp->router->aborted)
  4660. return 0;
  4661. vr = mlxsw_sp_vr_get(mlxsw_sp, ven_info->tb_id, NULL);
  4662. if (IS_ERR(vr))
  4663. return PTR_ERR(vr);
  4664. mrt = mlxsw_sp_router_fibmr_family_to_table(vr, ven_info->info.family);
  4665. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, ven_info->dev);
  4666. return mlxsw_sp_mr_vif_add(mrt, ven_info->dev,
  4667. ven_info->vif_index,
  4668. ven_info->vif_flags, rif);
  4669. }
  4670. static void
  4671. mlxsw_sp_router_fibmr_vif_del(struct mlxsw_sp *mlxsw_sp,
  4672. struct vif_entry_notifier_info *ven_info)
  4673. {
  4674. struct mlxsw_sp_mr_table *mrt;
  4675. struct mlxsw_sp_vr *vr;
  4676. if (mlxsw_sp->router->aborted)
  4677. return;
  4678. vr = mlxsw_sp_vr_find(mlxsw_sp, ven_info->tb_id);
  4679. if (WARN_ON(!vr))
  4680. return;
  4681. mrt = mlxsw_sp_router_fibmr_family_to_table(vr, ven_info->info.family);
  4682. mlxsw_sp_mr_vif_del(mrt, ven_info->vif_index);
  4683. mlxsw_sp_vr_put(mlxsw_sp, vr);
  4684. }
  4685. static int mlxsw_sp_router_set_abort_trap(struct mlxsw_sp *mlxsw_sp)
  4686. {
  4687. enum mlxsw_reg_ralxx_protocol proto = MLXSW_REG_RALXX_PROTOCOL_IPV4;
  4688. int err;
  4689. err = __mlxsw_sp_router_set_abort_trap(mlxsw_sp, proto,
  4690. MLXSW_SP_LPM_TREE_MIN);
  4691. if (err)
  4692. return err;
  4693. /* The multicast router code does not need an abort trap as by default,
  4694. * packets that don't match any routes are trapped to the CPU.
  4695. */
  4696. proto = MLXSW_REG_RALXX_PROTOCOL_IPV6;
  4697. return __mlxsw_sp_router_set_abort_trap(mlxsw_sp, proto,
  4698. MLXSW_SP_LPM_TREE_MIN + 1);
  4699. }
  4700. static void mlxsw_sp_fib4_node_flush(struct mlxsw_sp *mlxsw_sp,
  4701. struct mlxsw_sp_fib_node *fib_node)
  4702. {
  4703. struct mlxsw_sp_fib4_entry *fib4_entry, *tmp;
  4704. list_for_each_entry_safe(fib4_entry, tmp, &fib_node->entry_list,
  4705. common.list) {
  4706. bool do_break = &tmp->common.list == &fib_node->entry_list;
  4707. mlxsw_sp_fib4_node_entry_unlink(mlxsw_sp, fib4_entry);
  4708. mlxsw_sp_fib4_entry_destroy(mlxsw_sp, fib4_entry);
  4709. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4710. /* Break when entry list is empty and node was freed.
  4711. * Otherwise, we'll access freed memory in the next
  4712. * iteration.
  4713. */
  4714. if (do_break)
  4715. break;
  4716. }
  4717. }
  4718. static void mlxsw_sp_fib6_node_flush(struct mlxsw_sp *mlxsw_sp,
  4719. struct mlxsw_sp_fib_node *fib_node)
  4720. {
  4721. struct mlxsw_sp_fib6_entry *fib6_entry, *tmp;
  4722. list_for_each_entry_safe(fib6_entry, tmp, &fib_node->entry_list,
  4723. common.list) {
  4724. bool do_break = &tmp->common.list == &fib_node->entry_list;
  4725. mlxsw_sp_fib6_node_entry_unlink(mlxsw_sp, fib6_entry);
  4726. mlxsw_sp_fib6_entry_destroy(mlxsw_sp, fib6_entry);
  4727. mlxsw_sp_fib_node_put(mlxsw_sp, fib_node);
  4728. if (do_break)
  4729. break;
  4730. }
  4731. }
  4732. static void mlxsw_sp_fib_node_flush(struct mlxsw_sp *mlxsw_sp,
  4733. struct mlxsw_sp_fib_node *fib_node)
  4734. {
  4735. switch (fib_node->fib->proto) {
  4736. case MLXSW_SP_L3_PROTO_IPV4:
  4737. mlxsw_sp_fib4_node_flush(mlxsw_sp, fib_node);
  4738. break;
  4739. case MLXSW_SP_L3_PROTO_IPV6:
  4740. mlxsw_sp_fib6_node_flush(mlxsw_sp, fib_node);
  4741. break;
  4742. }
  4743. }
  4744. static void mlxsw_sp_vr_fib_flush(struct mlxsw_sp *mlxsw_sp,
  4745. struct mlxsw_sp_vr *vr,
  4746. enum mlxsw_sp_l3proto proto)
  4747. {
  4748. struct mlxsw_sp_fib *fib = mlxsw_sp_vr_fib(vr, proto);
  4749. struct mlxsw_sp_fib_node *fib_node, *tmp;
  4750. list_for_each_entry_safe(fib_node, tmp, &fib->node_list, list) {
  4751. bool do_break = &tmp->list == &fib->node_list;
  4752. mlxsw_sp_fib_node_flush(mlxsw_sp, fib_node);
  4753. if (do_break)
  4754. break;
  4755. }
  4756. }
  4757. static void mlxsw_sp_router_fib_flush(struct mlxsw_sp *mlxsw_sp)
  4758. {
  4759. int i, j;
  4760. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_VRS); i++) {
  4761. struct mlxsw_sp_vr *vr = &mlxsw_sp->router->vrs[i];
  4762. if (!mlxsw_sp_vr_is_used(vr))
  4763. continue;
  4764. for (j = 0; j < MLXSW_SP_L3_PROTO_MAX; j++)
  4765. mlxsw_sp_mr_table_flush(vr->mr_table[j]);
  4766. mlxsw_sp_vr_fib_flush(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV4);
  4767. /* If virtual router was only used for IPv4, then it's no
  4768. * longer used.
  4769. */
  4770. if (!mlxsw_sp_vr_is_used(vr))
  4771. continue;
  4772. mlxsw_sp_vr_fib_flush(mlxsw_sp, vr, MLXSW_SP_L3_PROTO_IPV6);
  4773. }
  4774. }
  4775. static void mlxsw_sp_router_fib_abort(struct mlxsw_sp *mlxsw_sp)
  4776. {
  4777. int err;
  4778. if (mlxsw_sp->router->aborted)
  4779. return;
  4780. dev_warn(mlxsw_sp->bus_info->dev, "FIB abort triggered. Note that FIB entries are no longer being offloaded to this device.\n");
  4781. mlxsw_sp_router_fib_flush(mlxsw_sp);
  4782. mlxsw_sp->router->aborted = true;
  4783. err = mlxsw_sp_router_set_abort_trap(mlxsw_sp);
  4784. if (err)
  4785. dev_warn(mlxsw_sp->bus_info->dev, "Failed to set abort trap.\n");
  4786. }
  4787. struct mlxsw_sp_fib_event_work {
  4788. struct work_struct work;
  4789. union {
  4790. struct fib6_entry_notifier_info fen6_info;
  4791. struct fib_entry_notifier_info fen_info;
  4792. struct fib_rule_notifier_info fr_info;
  4793. struct fib_nh_notifier_info fnh_info;
  4794. struct mfc_entry_notifier_info men_info;
  4795. struct vif_entry_notifier_info ven_info;
  4796. };
  4797. struct mlxsw_sp *mlxsw_sp;
  4798. unsigned long event;
  4799. };
  4800. static void mlxsw_sp_router_fib4_event_work(struct work_struct *work)
  4801. {
  4802. struct mlxsw_sp_fib_event_work *fib_work =
  4803. container_of(work, struct mlxsw_sp_fib_event_work, work);
  4804. struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
  4805. bool replace, append;
  4806. int err;
  4807. /* Protect internal structures from changes */
  4808. rtnl_lock();
  4809. mlxsw_sp_span_respin(mlxsw_sp);
  4810. switch (fib_work->event) {
  4811. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4812. case FIB_EVENT_ENTRY_APPEND: /* fall through */
  4813. case FIB_EVENT_ENTRY_ADD:
  4814. replace = fib_work->event == FIB_EVENT_ENTRY_REPLACE;
  4815. append = fib_work->event == FIB_EVENT_ENTRY_APPEND;
  4816. err = mlxsw_sp_router_fib4_add(mlxsw_sp, &fib_work->fen_info,
  4817. replace, append);
  4818. if (err)
  4819. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4820. fib_info_put(fib_work->fen_info.fi);
  4821. break;
  4822. case FIB_EVENT_ENTRY_DEL:
  4823. mlxsw_sp_router_fib4_del(mlxsw_sp, &fib_work->fen_info);
  4824. fib_info_put(fib_work->fen_info.fi);
  4825. break;
  4826. case FIB_EVENT_RULE_ADD:
  4827. /* if we get here, a rule was added that we do not support.
  4828. * just do the fib_abort
  4829. */
  4830. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4831. break;
  4832. case FIB_EVENT_NH_ADD: /* fall through */
  4833. case FIB_EVENT_NH_DEL:
  4834. mlxsw_sp_nexthop4_event(mlxsw_sp, fib_work->event,
  4835. fib_work->fnh_info.fib_nh);
  4836. fib_info_put(fib_work->fnh_info.fib_nh->nh_parent);
  4837. break;
  4838. }
  4839. rtnl_unlock();
  4840. kfree(fib_work);
  4841. }
  4842. static void mlxsw_sp_router_fib6_event_work(struct work_struct *work)
  4843. {
  4844. struct mlxsw_sp_fib_event_work *fib_work =
  4845. container_of(work, struct mlxsw_sp_fib_event_work, work);
  4846. struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
  4847. bool replace;
  4848. int err;
  4849. rtnl_lock();
  4850. mlxsw_sp_span_respin(mlxsw_sp);
  4851. switch (fib_work->event) {
  4852. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4853. case FIB_EVENT_ENTRY_APPEND: /* fall through */
  4854. case FIB_EVENT_ENTRY_ADD:
  4855. replace = fib_work->event == FIB_EVENT_ENTRY_REPLACE;
  4856. err = mlxsw_sp_router_fib6_add(mlxsw_sp,
  4857. fib_work->fen6_info.rt, replace);
  4858. if (err)
  4859. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4860. mlxsw_sp_rt6_release(fib_work->fen6_info.rt);
  4861. break;
  4862. case FIB_EVENT_ENTRY_DEL:
  4863. mlxsw_sp_router_fib6_del(mlxsw_sp, fib_work->fen6_info.rt);
  4864. mlxsw_sp_rt6_release(fib_work->fen6_info.rt);
  4865. break;
  4866. case FIB_EVENT_RULE_ADD:
  4867. /* if we get here, a rule was added that we do not support.
  4868. * just do the fib_abort
  4869. */
  4870. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4871. break;
  4872. }
  4873. rtnl_unlock();
  4874. kfree(fib_work);
  4875. }
  4876. static void mlxsw_sp_router_fibmr_event_work(struct work_struct *work)
  4877. {
  4878. struct mlxsw_sp_fib_event_work *fib_work =
  4879. container_of(work, struct mlxsw_sp_fib_event_work, work);
  4880. struct mlxsw_sp *mlxsw_sp = fib_work->mlxsw_sp;
  4881. bool replace;
  4882. int err;
  4883. rtnl_lock();
  4884. switch (fib_work->event) {
  4885. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4886. case FIB_EVENT_ENTRY_ADD:
  4887. replace = fib_work->event == FIB_EVENT_ENTRY_REPLACE;
  4888. err = mlxsw_sp_router_fibmr_add(mlxsw_sp, &fib_work->men_info,
  4889. replace);
  4890. if (err)
  4891. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4892. mr_cache_put(fib_work->men_info.mfc);
  4893. break;
  4894. case FIB_EVENT_ENTRY_DEL:
  4895. mlxsw_sp_router_fibmr_del(mlxsw_sp, &fib_work->men_info);
  4896. mr_cache_put(fib_work->men_info.mfc);
  4897. break;
  4898. case FIB_EVENT_VIF_ADD:
  4899. err = mlxsw_sp_router_fibmr_vif_add(mlxsw_sp,
  4900. &fib_work->ven_info);
  4901. if (err)
  4902. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4903. dev_put(fib_work->ven_info.dev);
  4904. break;
  4905. case FIB_EVENT_VIF_DEL:
  4906. mlxsw_sp_router_fibmr_vif_del(mlxsw_sp,
  4907. &fib_work->ven_info);
  4908. dev_put(fib_work->ven_info.dev);
  4909. break;
  4910. case FIB_EVENT_RULE_ADD:
  4911. /* if we get here, a rule was added that we do not support.
  4912. * just do the fib_abort
  4913. */
  4914. mlxsw_sp_router_fib_abort(mlxsw_sp);
  4915. break;
  4916. }
  4917. rtnl_unlock();
  4918. kfree(fib_work);
  4919. }
  4920. static void mlxsw_sp_router_fib4_event(struct mlxsw_sp_fib_event_work *fib_work,
  4921. struct fib_notifier_info *info)
  4922. {
  4923. struct fib_entry_notifier_info *fen_info;
  4924. struct fib_nh_notifier_info *fnh_info;
  4925. switch (fib_work->event) {
  4926. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4927. case FIB_EVENT_ENTRY_APPEND: /* fall through */
  4928. case FIB_EVENT_ENTRY_ADD: /* fall through */
  4929. case FIB_EVENT_ENTRY_DEL:
  4930. fen_info = container_of(info, struct fib_entry_notifier_info,
  4931. info);
  4932. fib_work->fen_info = *fen_info;
  4933. /* Take reference on fib_info to prevent it from being
  4934. * freed while work is queued. Release it afterwards.
  4935. */
  4936. fib_info_hold(fib_work->fen_info.fi);
  4937. break;
  4938. case FIB_EVENT_NH_ADD: /* fall through */
  4939. case FIB_EVENT_NH_DEL:
  4940. fnh_info = container_of(info, struct fib_nh_notifier_info,
  4941. info);
  4942. fib_work->fnh_info = *fnh_info;
  4943. fib_info_hold(fib_work->fnh_info.fib_nh->nh_parent);
  4944. break;
  4945. }
  4946. }
  4947. static void mlxsw_sp_router_fib6_event(struct mlxsw_sp_fib_event_work *fib_work,
  4948. struct fib_notifier_info *info)
  4949. {
  4950. struct fib6_entry_notifier_info *fen6_info;
  4951. switch (fib_work->event) {
  4952. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4953. case FIB_EVENT_ENTRY_APPEND: /* fall through */
  4954. case FIB_EVENT_ENTRY_ADD: /* fall through */
  4955. case FIB_EVENT_ENTRY_DEL:
  4956. fen6_info = container_of(info, struct fib6_entry_notifier_info,
  4957. info);
  4958. fib_work->fen6_info = *fen6_info;
  4959. fib6_info_hold(fib_work->fen6_info.rt);
  4960. break;
  4961. }
  4962. }
  4963. static void
  4964. mlxsw_sp_router_fibmr_event(struct mlxsw_sp_fib_event_work *fib_work,
  4965. struct fib_notifier_info *info)
  4966. {
  4967. switch (fib_work->event) {
  4968. case FIB_EVENT_ENTRY_REPLACE: /* fall through */
  4969. case FIB_EVENT_ENTRY_ADD: /* fall through */
  4970. case FIB_EVENT_ENTRY_DEL:
  4971. memcpy(&fib_work->men_info, info, sizeof(fib_work->men_info));
  4972. mr_cache_hold(fib_work->men_info.mfc);
  4973. break;
  4974. case FIB_EVENT_VIF_ADD: /* fall through */
  4975. case FIB_EVENT_VIF_DEL:
  4976. memcpy(&fib_work->ven_info, info, sizeof(fib_work->ven_info));
  4977. dev_hold(fib_work->ven_info.dev);
  4978. break;
  4979. }
  4980. }
  4981. static int mlxsw_sp_router_fib_rule_event(unsigned long event,
  4982. struct fib_notifier_info *info,
  4983. struct mlxsw_sp *mlxsw_sp)
  4984. {
  4985. struct netlink_ext_ack *extack = info->extack;
  4986. struct fib_rule_notifier_info *fr_info;
  4987. struct fib_rule *rule;
  4988. int err = 0;
  4989. /* nothing to do at the moment */
  4990. if (event == FIB_EVENT_RULE_DEL)
  4991. return 0;
  4992. if (mlxsw_sp->router->aborted)
  4993. return 0;
  4994. fr_info = container_of(info, struct fib_rule_notifier_info, info);
  4995. rule = fr_info->rule;
  4996. switch (info->family) {
  4997. case AF_INET:
  4998. if (!fib4_rule_default(rule) && !rule->l3mdev)
  4999. err = -EOPNOTSUPP;
  5000. break;
  5001. case AF_INET6:
  5002. if (!fib6_rule_default(rule) && !rule->l3mdev)
  5003. err = -EOPNOTSUPP;
  5004. break;
  5005. case RTNL_FAMILY_IPMR:
  5006. if (!ipmr_rule_default(rule) && !rule->l3mdev)
  5007. err = -EOPNOTSUPP;
  5008. break;
  5009. case RTNL_FAMILY_IP6MR:
  5010. if (!ip6mr_rule_default(rule) && !rule->l3mdev)
  5011. err = -EOPNOTSUPP;
  5012. break;
  5013. }
  5014. if (err < 0)
  5015. NL_SET_ERR_MSG_MOD(extack, "FIB rules not supported");
  5016. return err;
  5017. }
  5018. /* Called with rcu_read_lock() */
  5019. static int mlxsw_sp_router_fib_event(struct notifier_block *nb,
  5020. unsigned long event, void *ptr)
  5021. {
  5022. struct mlxsw_sp_fib_event_work *fib_work;
  5023. struct fib_notifier_info *info = ptr;
  5024. struct mlxsw_sp_router *router;
  5025. int err;
  5026. if (!net_eq(info->net, &init_net) ||
  5027. (info->family != AF_INET && info->family != AF_INET6 &&
  5028. info->family != RTNL_FAMILY_IPMR &&
  5029. info->family != RTNL_FAMILY_IP6MR))
  5030. return NOTIFY_DONE;
  5031. router = container_of(nb, struct mlxsw_sp_router, fib_nb);
  5032. switch (event) {
  5033. case FIB_EVENT_RULE_ADD: /* fall through */
  5034. case FIB_EVENT_RULE_DEL:
  5035. err = mlxsw_sp_router_fib_rule_event(event, info,
  5036. router->mlxsw_sp);
  5037. if (!err || info->extack)
  5038. return notifier_from_errno(err);
  5039. break;
  5040. case FIB_EVENT_ENTRY_ADD:
  5041. if (router->aborted) {
  5042. NL_SET_ERR_MSG_MOD(info->extack, "FIB offload was aborted. Not configuring route");
  5043. return notifier_from_errno(-EINVAL);
  5044. }
  5045. break;
  5046. }
  5047. fib_work = kzalloc(sizeof(*fib_work), GFP_ATOMIC);
  5048. if (WARN_ON(!fib_work))
  5049. return NOTIFY_BAD;
  5050. fib_work->mlxsw_sp = router->mlxsw_sp;
  5051. fib_work->event = event;
  5052. switch (info->family) {
  5053. case AF_INET:
  5054. INIT_WORK(&fib_work->work, mlxsw_sp_router_fib4_event_work);
  5055. mlxsw_sp_router_fib4_event(fib_work, info);
  5056. break;
  5057. case AF_INET6:
  5058. INIT_WORK(&fib_work->work, mlxsw_sp_router_fib6_event_work);
  5059. mlxsw_sp_router_fib6_event(fib_work, info);
  5060. break;
  5061. case RTNL_FAMILY_IP6MR:
  5062. case RTNL_FAMILY_IPMR:
  5063. INIT_WORK(&fib_work->work, mlxsw_sp_router_fibmr_event_work);
  5064. mlxsw_sp_router_fibmr_event(fib_work, info);
  5065. break;
  5066. }
  5067. mlxsw_core_schedule_work(&fib_work->work);
  5068. return NOTIFY_DONE;
  5069. }
  5070. static struct mlxsw_sp_rif *
  5071. mlxsw_sp_rif_find_by_dev(const struct mlxsw_sp *mlxsw_sp,
  5072. const struct net_device *dev)
  5073. {
  5074. int i;
  5075. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++)
  5076. if (mlxsw_sp->router->rifs[i] &&
  5077. mlxsw_sp->router->rifs[i]->dev == dev)
  5078. return mlxsw_sp->router->rifs[i];
  5079. return NULL;
  5080. }
  5081. static int mlxsw_sp_router_rif_disable(struct mlxsw_sp *mlxsw_sp, u16 rif)
  5082. {
  5083. char ritr_pl[MLXSW_REG_RITR_LEN];
  5084. int err;
  5085. mlxsw_reg_ritr_rif_pack(ritr_pl, rif);
  5086. err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5087. if (WARN_ON_ONCE(err))
  5088. return err;
  5089. mlxsw_reg_ritr_enable_set(ritr_pl, false);
  5090. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5091. }
  5092. static void mlxsw_sp_router_rif_gone_sync(struct mlxsw_sp *mlxsw_sp,
  5093. struct mlxsw_sp_rif *rif)
  5094. {
  5095. mlxsw_sp_router_rif_disable(mlxsw_sp, rif->rif_index);
  5096. mlxsw_sp_nexthop_rif_gone_sync(mlxsw_sp, rif);
  5097. mlxsw_sp_neigh_rif_gone_sync(mlxsw_sp, rif);
  5098. }
  5099. static bool
  5100. mlxsw_sp_rif_should_config(struct mlxsw_sp_rif *rif, struct net_device *dev,
  5101. unsigned long event)
  5102. {
  5103. struct inet6_dev *inet6_dev;
  5104. bool addr_list_empty = true;
  5105. struct in_device *idev;
  5106. switch (event) {
  5107. case NETDEV_UP:
  5108. return rif == NULL;
  5109. case NETDEV_DOWN:
  5110. idev = __in_dev_get_rtnl(dev);
  5111. if (idev && idev->ifa_list)
  5112. addr_list_empty = false;
  5113. inet6_dev = __in6_dev_get(dev);
  5114. if (addr_list_empty && inet6_dev &&
  5115. !list_empty(&inet6_dev->addr_list))
  5116. addr_list_empty = false;
  5117. if (rif && addr_list_empty &&
  5118. !netif_is_l3_slave(rif->dev))
  5119. return true;
  5120. /* It is possible we already removed the RIF ourselves
  5121. * if it was assigned to a netdev that is now a bridge
  5122. * or LAG slave.
  5123. */
  5124. return false;
  5125. }
  5126. return false;
  5127. }
  5128. static enum mlxsw_sp_rif_type
  5129. mlxsw_sp_dev_rif_type(const struct mlxsw_sp *mlxsw_sp,
  5130. const struct net_device *dev)
  5131. {
  5132. enum mlxsw_sp_fid_type type;
  5133. if (mlxsw_sp_netdev_ipip_type(mlxsw_sp, dev, NULL))
  5134. return MLXSW_SP_RIF_TYPE_IPIP_LB;
  5135. /* Otherwise RIF type is derived from the type of the underlying FID. */
  5136. if (is_vlan_dev(dev) && netif_is_bridge_master(vlan_dev_real_dev(dev)))
  5137. type = MLXSW_SP_FID_TYPE_8021Q;
  5138. else if (netif_is_bridge_master(dev) && br_vlan_enabled(dev))
  5139. type = MLXSW_SP_FID_TYPE_8021Q;
  5140. else if (netif_is_bridge_master(dev))
  5141. type = MLXSW_SP_FID_TYPE_8021D;
  5142. else
  5143. type = MLXSW_SP_FID_TYPE_RFID;
  5144. return mlxsw_sp_fid_type_rif_type(mlxsw_sp, type);
  5145. }
  5146. static int mlxsw_sp_rif_index_alloc(struct mlxsw_sp *mlxsw_sp, u16 *p_rif_index)
  5147. {
  5148. int i;
  5149. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++) {
  5150. if (!mlxsw_sp->router->rifs[i]) {
  5151. *p_rif_index = i;
  5152. return 0;
  5153. }
  5154. }
  5155. return -ENOBUFS;
  5156. }
  5157. static struct mlxsw_sp_rif *mlxsw_sp_rif_alloc(size_t rif_size, u16 rif_index,
  5158. u16 vr_id,
  5159. struct net_device *l3_dev)
  5160. {
  5161. struct mlxsw_sp_rif *rif;
  5162. rif = kzalloc(rif_size, GFP_KERNEL);
  5163. if (!rif)
  5164. return NULL;
  5165. INIT_LIST_HEAD(&rif->nexthop_list);
  5166. INIT_LIST_HEAD(&rif->neigh_list);
  5167. ether_addr_copy(rif->addr, l3_dev->dev_addr);
  5168. rif->mtu = l3_dev->mtu;
  5169. rif->vr_id = vr_id;
  5170. rif->dev = l3_dev;
  5171. rif->rif_index = rif_index;
  5172. return rif;
  5173. }
  5174. struct mlxsw_sp_rif *mlxsw_sp_rif_by_index(const struct mlxsw_sp *mlxsw_sp,
  5175. u16 rif_index)
  5176. {
  5177. return mlxsw_sp->router->rifs[rif_index];
  5178. }
  5179. u16 mlxsw_sp_rif_index(const struct mlxsw_sp_rif *rif)
  5180. {
  5181. return rif->rif_index;
  5182. }
  5183. u16 mlxsw_sp_ipip_lb_rif_index(const struct mlxsw_sp_rif_ipip_lb *lb_rif)
  5184. {
  5185. return lb_rif->common.rif_index;
  5186. }
  5187. u16 mlxsw_sp_ipip_lb_ul_vr_id(const struct mlxsw_sp_rif_ipip_lb *lb_rif)
  5188. {
  5189. return lb_rif->ul_vr_id;
  5190. }
  5191. int mlxsw_sp_rif_dev_ifindex(const struct mlxsw_sp_rif *rif)
  5192. {
  5193. return rif->dev->ifindex;
  5194. }
  5195. const struct net_device *mlxsw_sp_rif_dev(const struct mlxsw_sp_rif *rif)
  5196. {
  5197. return rif->dev;
  5198. }
  5199. static struct mlxsw_sp_rif *
  5200. mlxsw_sp_rif_create(struct mlxsw_sp *mlxsw_sp,
  5201. const struct mlxsw_sp_rif_params *params,
  5202. struct netlink_ext_ack *extack)
  5203. {
  5204. u32 tb_id = l3mdev_fib_table(params->dev);
  5205. const struct mlxsw_sp_rif_ops *ops;
  5206. struct mlxsw_sp_fid *fid = NULL;
  5207. enum mlxsw_sp_rif_type type;
  5208. struct mlxsw_sp_rif *rif;
  5209. struct mlxsw_sp_vr *vr;
  5210. u16 rif_index;
  5211. int i, err;
  5212. type = mlxsw_sp_dev_rif_type(mlxsw_sp, params->dev);
  5213. ops = mlxsw_sp->router->rif_ops_arr[type];
  5214. vr = mlxsw_sp_vr_get(mlxsw_sp, tb_id ? : RT_TABLE_MAIN, extack);
  5215. if (IS_ERR(vr))
  5216. return ERR_CAST(vr);
  5217. vr->rif_count++;
  5218. err = mlxsw_sp_rif_index_alloc(mlxsw_sp, &rif_index);
  5219. if (err) {
  5220. NL_SET_ERR_MSG_MOD(extack, "Exceeded number of supported router interfaces");
  5221. goto err_rif_index_alloc;
  5222. }
  5223. rif = mlxsw_sp_rif_alloc(ops->rif_size, rif_index, vr->id, params->dev);
  5224. if (!rif) {
  5225. err = -ENOMEM;
  5226. goto err_rif_alloc;
  5227. }
  5228. rif->mlxsw_sp = mlxsw_sp;
  5229. rif->ops = ops;
  5230. if (ops->fid_get) {
  5231. fid = ops->fid_get(rif);
  5232. if (IS_ERR(fid)) {
  5233. err = PTR_ERR(fid);
  5234. goto err_fid_get;
  5235. }
  5236. rif->fid = fid;
  5237. }
  5238. if (ops->setup)
  5239. ops->setup(rif, params);
  5240. err = ops->configure(rif);
  5241. if (err)
  5242. goto err_configure;
  5243. for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++) {
  5244. err = mlxsw_sp_mr_rif_add(vr->mr_table[i], rif);
  5245. if (err)
  5246. goto err_mr_rif_add;
  5247. }
  5248. mlxsw_sp_rif_counters_alloc(rif);
  5249. mlxsw_sp->router->rifs[rif_index] = rif;
  5250. return rif;
  5251. err_mr_rif_add:
  5252. for (i--; i >= 0; i--)
  5253. mlxsw_sp_mr_rif_del(vr->mr_table[i], rif);
  5254. ops->deconfigure(rif);
  5255. err_configure:
  5256. if (fid)
  5257. mlxsw_sp_fid_put(fid);
  5258. err_fid_get:
  5259. kfree(rif);
  5260. err_rif_alloc:
  5261. err_rif_index_alloc:
  5262. vr->rif_count--;
  5263. mlxsw_sp_vr_put(mlxsw_sp, vr);
  5264. return ERR_PTR(err);
  5265. }
  5266. void mlxsw_sp_rif_destroy(struct mlxsw_sp_rif *rif)
  5267. {
  5268. const struct mlxsw_sp_rif_ops *ops = rif->ops;
  5269. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5270. struct mlxsw_sp_fid *fid = rif->fid;
  5271. struct mlxsw_sp_vr *vr;
  5272. int i;
  5273. mlxsw_sp_router_rif_gone_sync(mlxsw_sp, rif);
  5274. vr = &mlxsw_sp->router->vrs[rif->vr_id];
  5275. mlxsw_sp->router->rifs[rif->rif_index] = NULL;
  5276. mlxsw_sp_rif_counters_free(rif);
  5277. for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++)
  5278. mlxsw_sp_mr_rif_del(vr->mr_table[i], rif);
  5279. ops->deconfigure(rif);
  5280. if (fid)
  5281. /* Loopback RIFs are not associated with a FID. */
  5282. mlxsw_sp_fid_put(fid);
  5283. kfree(rif);
  5284. vr->rif_count--;
  5285. mlxsw_sp_vr_put(mlxsw_sp, vr);
  5286. }
  5287. static void
  5288. mlxsw_sp_rif_subport_params_init(struct mlxsw_sp_rif_params *params,
  5289. struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan)
  5290. {
  5291. struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
  5292. params->vid = mlxsw_sp_port_vlan->vid;
  5293. params->lag = mlxsw_sp_port->lagged;
  5294. if (params->lag)
  5295. params->lag_id = mlxsw_sp_port->lag_id;
  5296. else
  5297. params->system_port = mlxsw_sp_port->local_port;
  5298. }
  5299. static int
  5300. mlxsw_sp_port_vlan_router_join(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan,
  5301. struct net_device *l3_dev,
  5302. struct netlink_ext_ack *extack)
  5303. {
  5304. struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
  5305. struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
  5306. u16 vid = mlxsw_sp_port_vlan->vid;
  5307. struct mlxsw_sp_rif *rif;
  5308. struct mlxsw_sp_fid *fid;
  5309. int err;
  5310. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
  5311. if (!rif) {
  5312. struct mlxsw_sp_rif_params params = {
  5313. .dev = l3_dev,
  5314. };
  5315. mlxsw_sp_rif_subport_params_init(&params, mlxsw_sp_port_vlan);
  5316. rif = mlxsw_sp_rif_create(mlxsw_sp, &params, extack);
  5317. if (IS_ERR(rif))
  5318. return PTR_ERR(rif);
  5319. }
  5320. /* FID was already created, just take a reference */
  5321. fid = rif->ops->fid_get(rif);
  5322. err = mlxsw_sp_fid_port_vid_map(fid, mlxsw_sp_port, vid);
  5323. if (err)
  5324. goto err_fid_port_vid_map;
  5325. err = mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, false);
  5326. if (err)
  5327. goto err_port_vid_learning_set;
  5328. err = mlxsw_sp_port_vid_stp_set(mlxsw_sp_port, vid,
  5329. BR_STATE_FORWARDING);
  5330. if (err)
  5331. goto err_port_vid_stp_set;
  5332. mlxsw_sp_port_vlan->fid = fid;
  5333. return 0;
  5334. err_port_vid_stp_set:
  5335. mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, true);
  5336. err_port_vid_learning_set:
  5337. mlxsw_sp_fid_port_vid_unmap(fid, mlxsw_sp_port, vid);
  5338. err_fid_port_vid_map:
  5339. mlxsw_sp_fid_put(fid);
  5340. return err;
  5341. }
  5342. void
  5343. mlxsw_sp_port_vlan_router_leave(struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan)
  5344. {
  5345. struct mlxsw_sp_port *mlxsw_sp_port = mlxsw_sp_port_vlan->mlxsw_sp_port;
  5346. struct mlxsw_sp_fid *fid = mlxsw_sp_port_vlan->fid;
  5347. u16 vid = mlxsw_sp_port_vlan->vid;
  5348. if (WARN_ON(mlxsw_sp_fid_type(fid) != MLXSW_SP_FID_TYPE_RFID))
  5349. return;
  5350. mlxsw_sp_port_vlan->fid = NULL;
  5351. mlxsw_sp_port_vid_stp_set(mlxsw_sp_port, vid, BR_STATE_BLOCKING);
  5352. mlxsw_sp_port_vid_learning_set(mlxsw_sp_port, vid, true);
  5353. mlxsw_sp_fid_port_vid_unmap(fid, mlxsw_sp_port, vid);
  5354. /* If router port holds the last reference on the rFID, then the
  5355. * associated Sub-port RIF will be destroyed.
  5356. */
  5357. mlxsw_sp_fid_put(fid);
  5358. }
  5359. static int mlxsw_sp_inetaddr_port_vlan_event(struct net_device *l3_dev,
  5360. struct net_device *port_dev,
  5361. unsigned long event, u16 vid,
  5362. struct netlink_ext_ack *extack)
  5363. {
  5364. struct mlxsw_sp_port *mlxsw_sp_port = netdev_priv(port_dev);
  5365. struct mlxsw_sp_port_vlan *mlxsw_sp_port_vlan;
  5366. mlxsw_sp_port_vlan = mlxsw_sp_port_vlan_find_by_vid(mlxsw_sp_port, vid);
  5367. if (WARN_ON(!mlxsw_sp_port_vlan))
  5368. return -EINVAL;
  5369. switch (event) {
  5370. case NETDEV_UP:
  5371. return mlxsw_sp_port_vlan_router_join(mlxsw_sp_port_vlan,
  5372. l3_dev, extack);
  5373. case NETDEV_DOWN:
  5374. mlxsw_sp_port_vlan_router_leave(mlxsw_sp_port_vlan);
  5375. break;
  5376. }
  5377. return 0;
  5378. }
  5379. static int mlxsw_sp_inetaddr_port_event(struct net_device *port_dev,
  5380. unsigned long event,
  5381. struct netlink_ext_ack *extack)
  5382. {
  5383. if (netif_is_bridge_port(port_dev) ||
  5384. netif_is_lag_port(port_dev) ||
  5385. netif_is_ovs_port(port_dev))
  5386. return 0;
  5387. return mlxsw_sp_inetaddr_port_vlan_event(port_dev, port_dev, event, 1,
  5388. extack);
  5389. }
  5390. static int __mlxsw_sp_inetaddr_lag_event(struct net_device *l3_dev,
  5391. struct net_device *lag_dev,
  5392. unsigned long event, u16 vid,
  5393. struct netlink_ext_ack *extack)
  5394. {
  5395. struct net_device *port_dev;
  5396. struct list_head *iter;
  5397. int err;
  5398. netdev_for_each_lower_dev(lag_dev, port_dev, iter) {
  5399. if (mlxsw_sp_port_dev_check(port_dev)) {
  5400. err = mlxsw_sp_inetaddr_port_vlan_event(l3_dev,
  5401. port_dev,
  5402. event, vid,
  5403. extack);
  5404. if (err)
  5405. return err;
  5406. }
  5407. }
  5408. return 0;
  5409. }
  5410. static int mlxsw_sp_inetaddr_lag_event(struct net_device *lag_dev,
  5411. unsigned long event,
  5412. struct netlink_ext_ack *extack)
  5413. {
  5414. if (netif_is_bridge_port(lag_dev))
  5415. return 0;
  5416. return __mlxsw_sp_inetaddr_lag_event(lag_dev, lag_dev, event, 1,
  5417. extack);
  5418. }
  5419. static int mlxsw_sp_inetaddr_bridge_event(struct net_device *l3_dev,
  5420. unsigned long event,
  5421. struct netlink_ext_ack *extack)
  5422. {
  5423. struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(l3_dev);
  5424. struct mlxsw_sp_rif_params params = {
  5425. .dev = l3_dev,
  5426. };
  5427. struct mlxsw_sp_rif *rif;
  5428. switch (event) {
  5429. case NETDEV_UP:
  5430. rif = mlxsw_sp_rif_create(mlxsw_sp, &params, extack);
  5431. if (IS_ERR(rif))
  5432. return PTR_ERR(rif);
  5433. break;
  5434. case NETDEV_DOWN:
  5435. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
  5436. mlxsw_sp_rif_destroy(rif);
  5437. break;
  5438. }
  5439. return 0;
  5440. }
  5441. static int mlxsw_sp_inetaddr_vlan_event(struct net_device *vlan_dev,
  5442. unsigned long event,
  5443. struct netlink_ext_ack *extack)
  5444. {
  5445. struct net_device *real_dev = vlan_dev_real_dev(vlan_dev);
  5446. u16 vid = vlan_dev_vlan_id(vlan_dev);
  5447. if (netif_is_bridge_port(vlan_dev))
  5448. return 0;
  5449. if (mlxsw_sp_port_dev_check(real_dev))
  5450. return mlxsw_sp_inetaddr_port_vlan_event(vlan_dev, real_dev,
  5451. event, vid, extack);
  5452. else if (netif_is_lag_master(real_dev))
  5453. return __mlxsw_sp_inetaddr_lag_event(vlan_dev, real_dev, event,
  5454. vid, extack);
  5455. else if (netif_is_bridge_master(real_dev) && br_vlan_enabled(real_dev))
  5456. return mlxsw_sp_inetaddr_bridge_event(vlan_dev, event, extack);
  5457. return 0;
  5458. }
  5459. static int __mlxsw_sp_inetaddr_event(struct net_device *dev,
  5460. unsigned long event,
  5461. struct netlink_ext_ack *extack)
  5462. {
  5463. if (mlxsw_sp_port_dev_check(dev))
  5464. return mlxsw_sp_inetaddr_port_event(dev, event, extack);
  5465. else if (netif_is_lag_master(dev))
  5466. return mlxsw_sp_inetaddr_lag_event(dev, event, extack);
  5467. else if (netif_is_bridge_master(dev))
  5468. return mlxsw_sp_inetaddr_bridge_event(dev, event, extack);
  5469. else if (is_vlan_dev(dev))
  5470. return mlxsw_sp_inetaddr_vlan_event(dev, event, extack);
  5471. else
  5472. return 0;
  5473. }
  5474. int mlxsw_sp_inetaddr_event(struct notifier_block *unused,
  5475. unsigned long event, void *ptr)
  5476. {
  5477. struct in_ifaddr *ifa = (struct in_ifaddr *) ptr;
  5478. struct net_device *dev = ifa->ifa_dev->dev;
  5479. struct mlxsw_sp *mlxsw_sp;
  5480. struct mlxsw_sp_rif *rif;
  5481. int err = 0;
  5482. /* NETDEV_UP event is handled by mlxsw_sp_inetaddr_valid_event */
  5483. if (event == NETDEV_UP)
  5484. goto out;
  5485. mlxsw_sp = mlxsw_sp_lower_get(dev);
  5486. if (!mlxsw_sp)
  5487. goto out;
  5488. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  5489. if (!mlxsw_sp_rif_should_config(rif, dev, event))
  5490. goto out;
  5491. err = __mlxsw_sp_inetaddr_event(dev, event, NULL);
  5492. out:
  5493. return notifier_from_errno(err);
  5494. }
  5495. int mlxsw_sp_inetaddr_valid_event(struct notifier_block *unused,
  5496. unsigned long event, void *ptr)
  5497. {
  5498. struct in_validator_info *ivi = (struct in_validator_info *) ptr;
  5499. struct net_device *dev = ivi->ivi_dev->dev;
  5500. struct mlxsw_sp *mlxsw_sp;
  5501. struct mlxsw_sp_rif *rif;
  5502. int err = 0;
  5503. mlxsw_sp = mlxsw_sp_lower_get(dev);
  5504. if (!mlxsw_sp)
  5505. goto out;
  5506. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  5507. if (!mlxsw_sp_rif_should_config(rif, dev, event))
  5508. goto out;
  5509. err = __mlxsw_sp_inetaddr_event(dev, event, ivi->extack);
  5510. out:
  5511. return notifier_from_errno(err);
  5512. }
  5513. struct mlxsw_sp_inet6addr_event_work {
  5514. struct work_struct work;
  5515. struct net_device *dev;
  5516. unsigned long event;
  5517. };
  5518. static void mlxsw_sp_inet6addr_event_work(struct work_struct *work)
  5519. {
  5520. struct mlxsw_sp_inet6addr_event_work *inet6addr_work =
  5521. container_of(work, struct mlxsw_sp_inet6addr_event_work, work);
  5522. struct net_device *dev = inet6addr_work->dev;
  5523. unsigned long event = inet6addr_work->event;
  5524. struct mlxsw_sp *mlxsw_sp;
  5525. struct mlxsw_sp_rif *rif;
  5526. rtnl_lock();
  5527. mlxsw_sp = mlxsw_sp_lower_get(dev);
  5528. if (!mlxsw_sp)
  5529. goto out;
  5530. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  5531. if (!mlxsw_sp_rif_should_config(rif, dev, event))
  5532. goto out;
  5533. __mlxsw_sp_inetaddr_event(dev, event, NULL);
  5534. out:
  5535. rtnl_unlock();
  5536. dev_put(dev);
  5537. kfree(inet6addr_work);
  5538. }
  5539. /* Called with rcu_read_lock() */
  5540. int mlxsw_sp_inet6addr_event(struct notifier_block *unused,
  5541. unsigned long event, void *ptr)
  5542. {
  5543. struct inet6_ifaddr *if6 = (struct inet6_ifaddr *) ptr;
  5544. struct mlxsw_sp_inet6addr_event_work *inet6addr_work;
  5545. struct net_device *dev = if6->idev->dev;
  5546. /* NETDEV_UP event is handled by mlxsw_sp_inet6addr_valid_event */
  5547. if (event == NETDEV_UP)
  5548. return NOTIFY_DONE;
  5549. if (!mlxsw_sp_port_dev_lower_find_rcu(dev))
  5550. return NOTIFY_DONE;
  5551. inet6addr_work = kzalloc(sizeof(*inet6addr_work), GFP_ATOMIC);
  5552. if (!inet6addr_work)
  5553. return NOTIFY_BAD;
  5554. INIT_WORK(&inet6addr_work->work, mlxsw_sp_inet6addr_event_work);
  5555. inet6addr_work->dev = dev;
  5556. inet6addr_work->event = event;
  5557. dev_hold(dev);
  5558. mlxsw_core_schedule_work(&inet6addr_work->work);
  5559. return NOTIFY_DONE;
  5560. }
  5561. int mlxsw_sp_inet6addr_valid_event(struct notifier_block *unused,
  5562. unsigned long event, void *ptr)
  5563. {
  5564. struct in6_validator_info *i6vi = (struct in6_validator_info *) ptr;
  5565. struct net_device *dev = i6vi->i6vi_dev->dev;
  5566. struct mlxsw_sp *mlxsw_sp;
  5567. struct mlxsw_sp_rif *rif;
  5568. int err = 0;
  5569. mlxsw_sp = mlxsw_sp_lower_get(dev);
  5570. if (!mlxsw_sp)
  5571. goto out;
  5572. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  5573. if (!mlxsw_sp_rif_should_config(rif, dev, event))
  5574. goto out;
  5575. err = __mlxsw_sp_inetaddr_event(dev, event, i6vi->extack);
  5576. out:
  5577. return notifier_from_errno(err);
  5578. }
  5579. static int mlxsw_sp_rif_edit(struct mlxsw_sp *mlxsw_sp, u16 rif_index,
  5580. const char *mac, int mtu)
  5581. {
  5582. char ritr_pl[MLXSW_REG_RITR_LEN];
  5583. int err;
  5584. mlxsw_reg_ritr_rif_pack(ritr_pl, rif_index);
  5585. err = mlxsw_reg_query(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5586. if (err)
  5587. return err;
  5588. mlxsw_reg_ritr_mtu_set(ritr_pl, mtu);
  5589. mlxsw_reg_ritr_if_mac_memcpy_to(ritr_pl, mac);
  5590. mlxsw_reg_ritr_op_set(ritr_pl, MLXSW_REG_RITR_RIF_CREATE);
  5591. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5592. }
  5593. int mlxsw_sp_netdevice_router_port_event(struct net_device *dev)
  5594. {
  5595. struct mlxsw_sp *mlxsw_sp;
  5596. struct mlxsw_sp_rif *rif;
  5597. u16 fid_index;
  5598. int err;
  5599. mlxsw_sp = mlxsw_sp_lower_get(dev);
  5600. if (!mlxsw_sp)
  5601. return 0;
  5602. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, dev);
  5603. if (!rif)
  5604. return 0;
  5605. fid_index = mlxsw_sp_fid_index(rif->fid);
  5606. err = mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, fid_index, false);
  5607. if (err)
  5608. return err;
  5609. err = mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, dev->dev_addr,
  5610. dev->mtu);
  5611. if (err)
  5612. goto err_rif_edit;
  5613. err = mlxsw_sp_rif_fdb_op(mlxsw_sp, dev->dev_addr, fid_index, true);
  5614. if (err)
  5615. goto err_rif_fdb_op;
  5616. if (rif->mtu != dev->mtu) {
  5617. struct mlxsw_sp_vr *vr;
  5618. int i;
  5619. /* The RIF is relevant only to its mr_table instance, as unlike
  5620. * unicast routing, in multicast routing a RIF cannot be shared
  5621. * between several multicast routing tables.
  5622. */
  5623. vr = &mlxsw_sp->router->vrs[rif->vr_id];
  5624. for (i = 0; i < MLXSW_SP_L3_PROTO_MAX; i++)
  5625. mlxsw_sp_mr_rif_mtu_update(vr->mr_table[i],
  5626. rif, dev->mtu);
  5627. }
  5628. ether_addr_copy(rif->addr, dev->dev_addr);
  5629. rif->mtu = dev->mtu;
  5630. netdev_dbg(dev, "Updated RIF=%d\n", rif->rif_index);
  5631. return 0;
  5632. err_rif_fdb_op:
  5633. mlxsw_sp_rif_edit(mlxsw_sp, rif->rif_index, rif->addr, rif->mtu);
  5634. err_rif_edit:
  5635. mlxsw_sp_rif_fdb_op(mlxsw_sp, rif->addr, fid_index, true);
  5636. return err;
  5637. }
  5638. static int mlxsw_sp_port_vrf_join(struct mlxsw_sp *mlxsw_sp,
  5639. struct net_device *l3_dev,
  5640. struct netlink_ext_ack *extack)
  5641. {
  5642. struct mlxsw_sp_rif *rif;
  5643. /* If netdev is already associated with a RIF, then we need to
  5644. * destroy it and create a new one with the new virtual router ID.
  5645. */
  5646. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
  5647. if (rif)
  5648. __mlxsw_sp_inetaddr_event(l3_dev, NETDEV_DOWN, extack);
  5649. return __mlxsw_sp_inetaddr_event(l3_dev, NETDEV_UP, extack);
  5650. }
  5651. static void mlxsw_sp_port_vrf_leave(struct mlxsw_sp *mlxsw_sp,
  5652. struct net_device *l3_dev)
  5653. {
  5654. struct mlxsw_sp_rif *rif;
  5655. rif = mlxsw_sp_rif_find_by_dev(mlxsw_sp, l3_dev);
  5656. if (!rif)
  5657. return;
  5658. __mlxsw_sp_inetaddr_event(l3_dev, NETDEV_DOWN, NULL);
  5659. }
  5660. int mlxsw_sp_netdevice_vrf_event(struct net_device *l3_dev, unsigned long event,
  5661. struct netdev_notifier_changeupper_info *info)
  5662. {
  5663. struct mlxsw_sp *mlxsw_sp = mlxsw_sp_lower_get(l3_dev);
  5664. int err = 0;
  5665. if (!mlxsw_sp)
  5666. return 0;
  5667. switch (event) {
  5668. case NETDEV_PRECHANGEUPPER:
  5669. return 0;
  5670. case NETDEV_CHANGEUPPER:
  5671. if (info->linking) {
  5672. struct netlink_ext_ack *extack;
  5673. extack = netdev_notifier_info_to_extack(&info->info);
  5674. err = mlxsw_sp_port_vrf_join(mlxsw_sp, l3_dev, extack);
  5675. } else {
  5676. mlxsw_sp_port_vrf_leave(mlxsw_sp, l3_dev);
  5677. }
  5678. break;
  5679. }
  5680. return err;
  5681. }
  5682. static struct mlxsw_sp_rif_subport *
  5683. mlxsw_sp_rif_subport_rif(const struct mlxsw_sp_rif *rif)
  5684. {
  5685. return container_of(rif, struct mlxsw_sp_rif_subport, common);
  5686. }
  5687. static void mlxsw_sp_rif_subport_setup(struct mlxsw_sp_rif *rif,
  5688. const struct mlxsw_sp_rif_params *params)
  5689. {
  5690. struct mlxsw_sp_rif_subport *rif_subport;
  5691. rif_subport = mlxsw_sp_rif_subport_rif(rif);
  5692. rif_subport->vid = params->vid;
  5693. rif_subport->lag = params->lag;
  5694. if (params->lag)
  5695. rif_subport->lag_id = params->lag_id;
  5696. else
  5697. rif_subport->system_port = params->system_port;
  5698. }
  5699. static int mlxsw_sp_rif_subport_op(struct mlxsw_sp_rif *rif, bool enable)
  5700. {
  5701. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5702. struct mlxsw_sp_rif_subport *rif_subport;
  5703. char ritr_pl[MLXSW_REG_RITR_LEN];
  5704. rif_subport = mlxsw_sp_rif_subport_rif(rif);
  5705. mlxsw_reg_ritr_pack(ritr_pl, enable, MLXSW_REG_RITR_SP_IF,
  5706. rif->rif_index, rif->vr_id, rif->dev->mtu);
  5707. mlxsw_reg_ritr_mac_pack(ritr_pl, rif->dev->dev_addr);
  5708. mlxsw_reg_ritr_sp_if_pack(ritr_pl, rif_subport->lag,
  5709. rif_subport->lag ? rif_subport->lag_id :
  5710. rif_subport->system_port,
  5711. rif_subport->vid);
  5712. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5713. }
  5714. static int mlxsw_sp_rif_subport_configure(struct mlxsw_sp_rif *rif)
  5715. {
  5716. int err;
  5717. err = mlxsw_sp_rif_subport_op(rif, true);
  5718. if (err)
  5719. return err;
  5720. err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5721. mlxsw_sp_fid_index(rif->fid), true);
  5722. if (err)
  5723. goto err_rif_fdb_op;
  5724. mlxsw_sp_fid_rif_set(rif->fid, rif);
  5725. return 0;
  5726. err_rif_fdb_op:
  5727. mlxsw_sp_rif_subport_op(rif, false);
  5728. return err;
  5729. }
  5730. static void mlxsw_sp_rif_subport_deconfigure(struct mlxsw_sp_rif *rif)
  5731. {
  5732. struct mlxsw_sp_fid *fid = rif->fid;
  5733. mlxsw_sp_fid_rif_set(fid, NULL);
  5734. mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5735. mlxsw_sp_fid_index(fid), false);
  5736. mlxsw_sp_rif_subport_op(rif, false);
  5737. }
  5738. static struct mlxsw_sp_fid *
  5739. mlxsw_sp_rif_subport_fid_get(struct mlxsw_sp_rif *rif)
  5740. {
  5741. return mlxsw_sp_fid_rfid_get(rif->mlxsw_sp, rif->rif_index);
  5742. }
  5743. static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_subport_ops = {
  5744. .type = MLXSW_SP_RIF_TYPE_SUBPORT,
  5745. .rif_size = sizeof(struct mlxsw_sp_rif_subport),
  5746. .setup = mlxsw_sp_rif_subport_setup,
  5747. .configure = mlxsw_sp_rif_subport_configure,
  5748. .deconfigure = mlxsw_sp_rif_subport_deconfigure,
  5749. .fid_get = mlxsw_sp_rif_subport_fid_get,
  5750. };
  5751. static int mlxsw_sp_rif_vlan_fid_op(struct mlxsw_sp_rif *rif,
  5752. enum mlxsw_reg_ritr_if_type type,
  5753. u16 vid_fid, bool enable)
  5754. {
  5755. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5756. char ritr_pl[MLXSW_REG_RITR_LEN];
  5757. mlxsw_reg_ritr_pack(ritr_pl, enable, type, rif->rif_index, rif->vr_id,
  5758. rif->dev->mtu);
  5759. mlxsw_reg_ritr_mac_pack(ritr_pl, rif->dev->dev_addr);
  5760. mlxsw_reg_ritr_fid_set(ritr_pl, type, vid_fid);
  5761. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ritr), ritr_pl);
  5762. }
  5763. u8 mlxsw_sp_router_port(const struct mlxsw_sp *mlxsw_sp)
  5764. {
  5765. return mlxsw_core_max_ports(mlxsw_sp->core) + 1;
  5766. }
  5767. static int mlxsw_sp_rif_vlan_configure(struct mlxsw_sp_rif *rif)
  5768. {
  5769. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5770. u16 vid = mlxsw_sp_fid_8021q_vid(rif->fid);
  5771. int err;
  5772. err = mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_VLAN_IF, vid, true);
  5773. if (err)
  5774. return err;
  5775. err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5776. mlxsw_sp_router_port(mlxsw_sp), true);
  5777. if (err)
  5778. goto err_fid_mc_flood_set;
  5779. err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5780. mlxsw_sp_router_port(mlxsw_sp), true);
  5781. if (err)
  5782. goto err_fid_bc_flood_set;
  5783. err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5784. mlxsw_sp_fid_index(rif->fid), true);
  5785. if (err)
  5786. goto err_rif_fdb_op;
  5787. mlxsw_sp_fid_rif_set(rif->fid, rif);
  5788. return 0;
  5789. err_rif_fdb_op:
  5790. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5791. mlxsw_sp_router_port(mlxsw_sp), false);
  5792. err_fid_bc_flood_set:
  5793. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5794. mlxsw_sp_router_port(mlxsw_sp), false);
  5795. err_fid_mc_flood_set:
  5796. mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_VLAN_IF, vid, false);
  5797. return err;
  5798. }
  5799. static void mlxsw_sp_rif_vlan_deconfigure(struct mlxsw_sp_rif *rif)
  5800. {
  5801. u16 vid = mlxsw_sp_fid_8021q_vid(rif->fid);
  5802. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5803. struct mlxsw_sp_fid *fid = rif->fid;
  5804. mlxsw_sp_fid_rif_set(fid, NULL);
  5805. mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5806. mlxsw_sp_fid_index(fid), false);
  5807. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5808. mlxsw_sp_router_port(mlxsw_sp), false);
  5809. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5810. mlxsw_sp_router_port(mlxsw_sp), false);
  5811. mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_VLAN_IF, vid, false);
  5812. }
  5813. static struct mlxsw_sp_fid *
  5814. mlxsw_sp_rif_vlan_fid_get(struct mlxsw_sp_rif *rif)
  5815. {
  5816. u16 vid = is_vlan_dev(rif->dev) ? vlan_dev_vlan_id(rif->dev) : 1;
  5817. return mlxsw_sp_fid_8021q_get(rif->mlxsw_sp, vid);
  5818. }
  5819. static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_vlan_ops = {
  5820. .type = MLXSW_SP_RIF_TYPE_VLAN,
  5821. .rif_size = sizeof(struct mlxsw_sp_rif),
  5822. .configure = mlxsw_sp_rif_vlan_configure,
  5823. .deconfigure = mlxsw_sp_rif_vlan_deconfigure,
  5824. .fid_get = mlxsw_sp_rif_vlan_fid_get,
  5825. };
  5826. static int mlxsw_sp_rif_fid_configure(struct mlxsw_sp_rif *rif)
  5827. {
  5828. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5829. u16 fid_index = mlxsw_sp_fid_index(rif->fid);
  5830. int err;
  5831. err = mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_FID_IF, fid_index,
  5832. true);
  5833. if (err)
  5834. return err;
  5835. err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5836. mlxsw_sp_router_port(mlxsw_sp), true);
  5837. if (err)
  5838. goto err_fid_mc_flood_set;
  5839. err = mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5840. mlxsw_sp_router_port(mlxsw_sp), true);
  5841. if (err)
  5842. goto err_fid_bc_flood_set;
  5843. err = mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5844. mlxsw_sp_fid_index(rif->fid), true);
  5845. if (err)
  5846. goto err_rif_fdb_op;
  5847. mlxsw_sp_fid_rif_set(rif->fid, rif);
  5848. return 0;
  5849. err_rif_fdb_op:
  5850. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5851. mlxsw_sp_router_port(mlxsw_sp), false);
  5852. err_fid_bc_flood_set:
  5853. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5854. mlxsw_sp_router_port(mlxsw_sp), false);
  5855. err_fid_mc_flood_set:
  5856. mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_FID_IF, fid_index, false);
  5857. return err;
  5858. }
  5859. static void mlxsw_sp_rif_fid_deconfigure(struct mlxsw_sp_rif *rif)
  5860. {
  5861. u16 fid_index = mlxsw_sp_fid_index(rif->fid);
  5862. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5863. struct mlxsw_sp_fid *fid = rif->fid;
  5864. mlxsw_sp_fid_rif_set(fid, NULL);
  5865. mlxsw_sp_rif_fdb_op(rif->mlxsw_sp, rif->dev->dev_addr,
  5866. mlxsw_sp_fid_index(fid), false);
  5867. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_BC,
  5868. mlxsw_sp_router_port(mlxsw_sp), false);
  5869. mlxsw_sp_fid_flood_set(rif->fid, MLXSW_SP_FLOOD_TYPE_MC,
  5870. mlxsw_sp_router_port(mlxsw_sp), false);
  5871. mlxsw_sp_rif_vlan_fid_op(rif, MLXSW_REG_RITR_FID_IF, fid_index, false);
  5872. }
  5873. static struct mlxsw_sp_fid *
  5874. mlxsw_sp_rif_fid_fid_get(struct mlxsw_sp_rif *rif)
  5875. {
  5876. return mlxsw_sp_fid_8021d_get(rif->mlxsw_sp, rif->dev->ifindex);
  5877. }
  5878. static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_fid_ops = {
  5879. .type = MLXSW_SP_RIF_TYPE_FID,
  5880. .rif_size = sizeof(struct mlxsw_sp_rif),
  5881. .configure = mlxsw_sp_rif_fid_configure,
  5882. .deconfigure = mlxsw_sp_rif_fid_deconfigure,
  5883. .fid_get = mlxsw_sp_rif_fid_fid_get,
  5884. };
  5885. static struct mlxsw_sp_rif_ipip_lb *
  5886. mlxsw_sp_rif_ipip_lb_rif(struct mlxsw_sp_rif *rif)
  5887. {
  5888. return container_of(rif, struct mlxsw_sp_rif_ipip_lb, common);
  5889. }
  5890. static void
  5891. mlxsw_sp_rif_ipip_lb_setup(struct mlxsw_sp_rif *rif,
  5892. const struct mlxsw_sp_rif_params *params)
  5893. {
  5894. struct mlxsw_sp_rif_params_ipip_lb *params_lb;
  5895. struct mlxsw_sp_rif_ipip_lb *rif_lb;
  5896. params_lb = container_of(params, struct mlxsw_sp_rif_params_ipip_lb,
  5897. common);
  5898. rif_lb = mlxsw_sp_rif_ipip_lb_rif(rif);
  5899. rif_lb->lb_config = params_lb->lb_config;
  5900. }
  5901. static int
  5902. mlxsw_sp_rif_ipip_lb_configure(struct mlxsw_sp_rif *rif)
  5903. {
  5904. struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
  5905. u32 ul_tb_id = mlxsw_sp_ipip_dev_ul_tb_id(rif->dev);
  5906. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5907. struct mlxsw_sp_vr *ul_vr;
  5908. int err;
  5909. ul_vr = mlxsw_sp_vr_get(mlxsw_sp, ul_tb_id, NULL);
  5910. if (IS_ERR(ul_vr))
  5911. return PTR_ERR(ul_vr);
  5912. err = mlxsw_sp_rif_ipip_lb_op(lb_rif, ul_vr, true);
  5913. if (err)
  5914. goto err_loopback_op;
  5915. lb_rif->ul_vr_id = ul_vr->id;
  5916. ++ul_vr->rif_count;
  5917. return 0;
  5918. err_loopback_op:
  5919. mlxsw_sp_vr_put(mlxsw_sp, ul_vr);
  5920. return err;
  5921. }
  5922. static void mlxsw_sp_rif_ipip_lb_deconfigure(struct mlxsw_sp_rif *rif)
  5923. {
  5924. struct mlxsw_sp_rif_ipip_lb *lb_rif = mlxsw_sp_rif_ipip_lb_rif(rif);
  5925. struct mlxsw_sp *mlxsw_sp = rif->mlxsw_sp;
  5926. struct mlxsw_sp_vr *ul_vr;
  5927. ul_vr = &mlxsw_sp->router->vrs[lb_rif->ul_vr_id];
  5928. mlxsw_sp_rif_ipip_lb_op(lb_rif, ul_vr, false);
  5929. --ul_vr->rif_count;
  5930. mlxsw_sp_vr_put(mlxsw_sp, ul_vr);
  5931. }
  5932. static const struct mlxsw_sp_rif_ops mlxsw_sp_rif_ipip_lb_ops = {
  5933. .type = MLXSW_SP_RIF_TYPE_IPIP_LB,
  5934. .rif_size = sizeof(struct mlxsw_sp_rif_ipip_lb),
  5935. .setup = mlxsw_sp_rif_ipip_lb_setup,
  5936. .configure = mlxsw_sp_rif_ipip_lb_configure,
  5937. .deconfigure = mlxsw_sp_rif_ipip_lb_deconfigure,
  5938. };
  5939. static const struct mlxsw_sp_rif_ops *mlxsw_sp_rif_ops_arr[] = {
  5940. [MLXSW_SP_RIF_TYPE_SUBPORT] = &mlxsw_sp_rif_subport_ops,
  5941. [MLXSW_SP_RIF_TYPE_VLAN] = &mlxsw_sp_rif_vlan_ops,
  5942. [MLXSW_SP_RIF_TYPE_FID] = &mlxsw_sp_rif_fid_ops,
  5943. [MLXSW_SP_RIF_TYPE_IPIP_LB] = &mlxsw_sp_rif_ipip_lb_ops,
  5944. };
  5945. static int mlxsw_sp_rifs_init(struct mlxsw_sp *mlxsw_sp)
  5946. {
  5947. u64 max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
  5948. mlxsw_sp->router->rifs = kcalloc(max_rifs,
  5949. sizeof(struct mlxsw_sp_rif *),
  5950. GFP_KERNEL);
  5951. if (!mlxsw_sp->router->rifs)
  5952. return -ENOMEM;
  5953. mlxsw_sp->router->rif_ops_arr = mlxsw_sp_rif_ops_arr;
  5954. return 0;
  5955. }
  5956. static void mlxsw_sp_rifs_fini(struct mlxsw_sp *mlxsw_sp)
  5957. {
  5958. int i;
  5959. for (i = 0; i < MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS); i++)
  5960. WARN_ON_ONCE(mlxsw_sp->router->rifs[i]);
  5961. kfree(mlxsw_sp->router->rifs);
  5962. }
  5963. static int
  5964. mlxsw_sp_ipip_config_tigcr(struct mlxsw_sp *mlxsw_sp)
  5965. {
  5966. char tigcr_pl[MLXSW_REG_TIGCR_LEN];
  5967. mlxsw_reg_tigcr_pack(tigcr_pl, true, 0);
  5968. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(tigcr), tigcr_pl);
  5969. }
  5970. static int mlxsw_sp_ipips_init(struct mlxsw_sp *mlxsw_sp)
  5971. {
  5972. mlxsw_sp->router->ipip_ops_arr = mlxsw_sp_ipip_ops_arr;
  5973. INIT_LIST_HEAD(&mlxsw_sp->router->ipip_list);
  5974. return mlxsw_sp_ipip_config_tigcr(mlxsw_sp);
  5975. }
  5976. static void mlxsw_sp_ipips_fini(struct mlxsw_sp *mlxsw_sp)
  5977. {
  5978. WARN_ON(!list_empty(&mlxsw_sp->router->ipip_list));
  5979. }
  5980. static void mlxsw_sp_router_fib_dump_flush(struct notifier_block *nb)
  5981. {
  5982. struct mlxsw_sp_router *router;
  5983. /* Flush pending FIB notifications and then flush the device's
  5984. * table before requesting another dump. The FIB notification
  5985. * block is unregistered, so no need to take RTNL.
  5986. */
  5987. mlxsw_core_flush_owq();
  5988. router = container_of(nb, struct mlxsw_sp_router, fib_nb);
  5989. mlxsw_sp_router_fib_flush(router->mlxsw_sp);
  5990. }
  5991. #ifdef CONFIG_IP_ROUTE_MULTIPATH
  5992. static void mlxsw_sp_mp_hash_header_set(char *recr2_pl, int header)
  5993. {
  5994. mlxsw_reg_recr2_outer_header_enables_set(recr2_pl, header, true);
  5995. }
  5996. static void mlxsw_sp_mp_hash_field_set(char *recr2_pl, int field)
  5997. {
  5998. mlxsw_reg_recr2_outer_header_fields_enable_set(recr2_pl, field, true);
  5999. }
  6000. static void mlxsw_sp_mp4_hash_init(char *recr2_pl)
  6001. {
  6002. bool only_l3 = !init_net.ipv4.sysctl_fib_multipath_hash_policy;
  6003. mlxsw_sp_mp_hash_header_set(recr2_pl,
  6004. MLXSW_REG_RECR2_IPV4_EN_NOT_TCP_NOT_UDP);
  6005. mlxsw_sp_mp_hash_header_set(recr2_pl, MLXSW_REG_RECR2_IPV4_EN_TCP_UDP);
  6006. mlxsw_reg_recr2_ipv4_sip_enable(recr2_pl);
  6007. mlxsw_reg_recr2_ipv4_dip_enable(recr2_pl);
  6008. if (only_l3)
  6009. return;
  6010. mlxsw_sp_mp_hash_header_set(recr2_pl, MLXSW_REG_RECR2_TCP_UDP_EN_IPV4);
  6011. mlxsw_sp_mp_hash_field_set(recr2_pl, MLXSW_REG_RECR2_IPV4_PROTOCOL);
  6012. mlxsw_sp_mp_hash_field_set(recr2_pl, MLXSW_REG_RECR2_TCP_UDP_SPORT);
  6013. mlxsw_sp_mp_hash_field_set(recr2_pl, MLXSW_REG_RECR2_TCP_UDP_DPORT);
  6014. }
  6015. static void mlxsw_sp_mp6_hash_init(char *recr2_pl)
  6016. {
  6017. bool only_l3 = !ip6_multipath_hash_policy(&init_net);
  6018. mlxsw_sp_mp_hash_header_set(recr2_pl,
  6019. MLXSW_REG_RECR2_IPV6_EN_NOT_TCP_NOT_UDP);
  6020. mlxsw_sp_mp_hash_header_set(recr2_pl, MLXSW_REG_RECR2_IPV6_EN_TCP_UDP);
  6021. mlxsw_reg_recr2_ipv6_sip_enable(recr2_pl);
  6022. mlxsw_reg_recr2_ipv6_dip_enable(recr2_pl);
  6023. mlxsw_sp_mp_hash_field_set(recr2_pl, MLXSW_REG_RECR2_IPV6_NEXT_HEADER);
  6024. if (only_l3) {
  6025. mlxsw_sp_mp_hash_field_set(recr2_pl,
  6026. MLXSW_REG_RECR2_IPV6_FLOW_LABEL);
  6027. } else {
  6028. mlxsw_sp_mp_hash_header_set(recr2_pl,
  6029. MLXSW_REG_RECR2_TCP_UDP_EN_IPV6);
  6030. mlxsw_sp_mp_hash_field_set(recr2_pl,
  6031. MLXSW_REG_RECR2_TCP_UDP_SPORT);
  6032. mlxsw_sp_mp_hash_field_set(recr2_pl,
  6033. MLXSW_REG_RECR2_TCP_UDP_DPORT);
  6034. }
  6035. }
  6036. static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp)
  6037. {
  6038. char recr2_pl[MLXSW_REG_RECR2_LEN];
  6039. u32 seed;
  6040. get_random_bytes(&seed, sizeof(seed));
  6041. mlxsw_reg_recr2_pack(recr2_pl, seed);
  6042. mlxsw_sp_mp4_hash_init(recr2_pl);
  6043. mlxsw_sp_mp6_hash_init(recr2_pl);
  6044. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(recr2), recr2_pl);
  6045. }
  6046. #else
  6047. static int mlxsw_sp_mp_hash_init(struct mlxsw_sp *mlxsw_sp)
  6048. {
  6049. return 0;
  6050. }
  6051. #endif
  6052. static int mlxsw_sp_dscp_init(struct mlxsw_sp *mlxsw_sp)
  6053. {
  6054. char rdpm_pl[MLXSW_REG_RDPM_LEN];
  6055. unsigned int i;
  6056. MLXSW_REG_ZERO(rdpm, rdpm_pl);
  6057. /* HW is determining switch priority based on DSCP-bits, but the
  6058. * kernel is still doing that based on the ToS. Since there's a
  6059. * mismatch in bits we need to make sure to translate the right
  6060. * value ToS would observe, skipping the 2 least-significant ECN bits.
  6061. */
  6062. for (i = 0; i < MLXSW_REG_RDPM_DSCP_ENTRY_REC_MAX_COUNT; i++)
  6063. mlxsw_reg_rdpm_pack(rdpm_pl, i, rt_tos2priority(i << 2));
  6064. return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rdpm), rdpm_pl);
  6065. }
  6066. static int __mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
  6067. {
  6068. char rgcr_pl[MLXSW_REG_RGCR_LEN];
  6069. u64 max_rifs;
  6070. int err;
  6071. if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MAX_RIFS))
  6072. return -EIO;
  6073. max_rifs = MLXSW_CORE_RES_GET(mlxsw_sp->core, MAX_RIFS);
  6074. mlxsw_reg_rgcr_pack(rgcr_pl, true, true);
  6075. mlxsw_reg_rgcr_max_router_interfaces_set(rgcr_pl, max_rifs);
  6076. mlxsw_reg_rgcr_usp_set(rgcr_pl, true);
  6077. err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);
  6078. if (err)
  6079. return err;
  6080. return 0;
  6081. }
  6082. static void __mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
  6083. {
  6084. char rgcr_pl[MLXSW_REG_RGCR_LEN];
  6085. mlxsw_reg_rgcr_pack(rgcr_pl, false, false);
  6086. mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rgcr), rgcr_pl);
  6087. }
  6088. int mlxsw_sp_router_init(struct mlxsw_sp *mlxsw_sp)
  6089. {
  6090. struct mlxsw_sp_router *router;
  6091. int err;
  6092. router = kzalloc(sizeof(*mlxsw_sp->router), GFP_KERNEL);
  6093. if (!router)
  6094. return -ENOMEM;
  6095. mlxsw_sp->router = router;
  6096. router->mlxsw_sp = mlxsw_sp;
  6097. INIT_LIST_HEAD(&mlxsw_sp->router->nexthop_neighs_list);
  6098. err = __mlxsw_sp_router_init(mlxsw_sp);
  6099. if (err)
  6100. goto err_router_init;
  6101. err = mlxsw_sp_rifs_init(mlxsw_sp);
  6102. if (err)
  6103. goto err_rifs_init;
  6104. err = mlxsw_sp_ipips_init(mlxsw_sp);
  6105. if (err)
  6106. goto err_ipips_init;
  6107. err = rhashtable_init(&mlxsw_sp->router->nexthop_ht,
  6108. &mlxsw_sp_nexthop_ht_params);
  6109. if (err)
  6110. goto err_nexthop_ht_init;
  6111. err = rhashtable_init(&mlxsw_sp->router->nexthop_group_ht,
  6112. &mlxsw_sp_nexthop_group_ht_params);
  6113. if (err)
  6114. goto err_nexthop_group_ht_init;
  6115. INIT_LIST_HEAD(&mlxsw_sp->router->nexthop_list);
  6116. err = mlxsw_sp_lpm_init(mlxsw_sp);
  6117. if (err)
  6118. goto err_lpm_init;
  6119. err = mlxsw_sp_mr_init(mlxsw_sp, &mlxsw_sp_mr_tcam_ops);
  6120. if (err)
  6121. goto err_mr_init;
  6122. err = mlxsw_sp_vrs_init(mlxsw_sp);
  6123. if (err)
  6124. goto err_vrs_init;
  6125. err = mlxsw_sp_neigh_init(mlxsw_sp);
  6126. if (err)
  6127. goto err_neigh_init;
  6128. mlxsw_sp->router->netevent_nb.notifier_call =
  6129. mlxsw_sp_router_netevent_event;
  6130. err = register_netevent_notifier(&mlxsw_sp->router->netevent_nb);
  6131. if (err)
  6132. goto err_register_netevent_notifier;
  6133. err = mlxsw_sp_mp_hash_init(mlxsw_sp);
  6134. if (err)
  6135. goto err_mp_hash_init;
  6136. err = mlxsw_sp_dscp_init(mlxsw_sp);
  6137. if (err)
  6138. goto err_dscp_init;
  6139. mlxsw_sp->router->fib_nb.notifier_call = mlxsw_sp_router_fib_event;
  6140. err = register_fib_notifier(&mlxsw_sp->router->fib_nb,
  6141. mlxsw_sp_router_fib_dump_flush);
  6142. if (err)
  6143. goto err_register_fib_notifier;
  6144. return 0;
  6145. err_register_fib_notifier:
  6146. err_dscp_init:
  6147. err_mp_hash_init:
  6148. unregister_netevent_notifier(&mlxsw_sp->router->netevent_nb);
  6149. err_register_netevent_notifier:
  6150. mlxsw_sp_neigh_fini(mlxsw_sp);
  6151. err_neigh_init:
  6152. mlxsw_sp_vrs_fini(mlxsw_sp);
  6153. err_vrs_init:
  6154. mlxsw_sp_mr_fini(mlxsw_sp);
  6155. err_mr_init:
  6156. mlxsw_sp_lpm_fini(mlxsw_sp);
  6157. err_lpm_init:
  6158. rhashtable_destroy(&mlxsw_sp->router->nexthop_group_ht);
  6159. err_nexthop_group_ht_init:
  6160. rhashtable_destroy(&mlxsw_sp->router->nexthop_ht);
  6161. err_nexthop_ht_init:
  6162. mlxsw_sp_ipips_fini(mlxsw_sp);
  6163. err_ipips_init:
  6164. mlxsw_sp_rifs_fini(mlxsw_sp);
  6165. err_rifs_init:
  6166. __mlxsw_sp_router_fini(mlxsw_sp);
  6167. err_router_init:
  6168. kfree(mlxsw_sp->router);
  6169. return err;
  6170. }
  6171. void mlxsw_sp_router_fini(struct mlxsw_sp *mlxsw_sp)
  6172. {
  6173. unregister_fib_notifier(&mlxsw_sp->router->fib_nb);
  6174. unregister_netevent_notifier(&mlxsw_sp->router->netevent_nb);
  6175. mlxsw_sp_neigh_fini(mlxsw_sp);
  6176. mlxsw_sp_vrs_fini(mlxsw_sp);
  6177. mlxsw_sp_mr_fini(mlxsw_sp);
  6178. mlxsw_sp_lpm_fini(mlxsw_sp);
  6179. rhashtable_destroy(&mlxsw_sp->router->nexthop_group_ht);
  6180. rhashtable_destroy(&mlxsw_sp->router->nexthop_ht);
  6181. mlxsw_sp_ipips_fini(mlxsw_sp);
  6182. mlxsw_sp_rifs_fini(mlxsw_sp);
  6183. __mlxsw_sp_router_fini(mlxsw_sp);
  6184. kfree(mlxsw_sp->router);
  6185. }