forked from mirrors/linux
		
	net: reformat kdoc return statements
kernel-doc -Wall warns about missing Return: statement for non-void functions. We have a number of kdocs in our headers which are missing the colon, IOW they use * Return some value or * Returns some value Having the colon makes some sense, it should help kdoc parser avoid false positives. So add them. This is mostly done with a sed script, and removing the unnecessary cases (mostly the comments which aren't kdoc). Acked-by: Johannes Berg <johannes@sipsolutions.net> Acked-by: Richard Cochran <richardcochran@gmail.com> Acked-by: Sergey Ryazanov <ryazanov.s.a@gmail.com> Reviewed-by: Edward Cree <ecree.xilinx@gmail.com> Acked-by: Alexandra Winter <wintera@linux.ibm.com> Acked-by: Pablo Neira Ayuso <pablo@netfilter.org> Link: https://patch.msgid.link/20241205165914.1071102-1-kuba@kernel.org Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
		
							parent
							
								
									ca78588805
								
							
						
					
					
						commit
						3f330db306
					
				
					 21 changed files with 101 additions and 101 deletions
				
			
		|  | @ -81,7 +81,7 @@ static const u8 eth_ipv6_mcast_addr_base[ETH_ALEN] __aligned(2) = | |||
|  * is_link_local_ether_addr - Determine if given Ethernet address is link-local | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if address is link local reserved addr (01:80:c2:00:00:0X) per | ||||
|  * Return: true if address is link local reserved addr (01:80:c2:00:00:0X) per | ||||
|  * IEEE 802.1Q 8.6.3 Frame filtering. | ||||
|  * | ||||
|  * Please note: addr must be aligned to u16. | ||||
|  | @ -104,7 +104,7 @@ static inline bool is_link_local_ether_addr(const u8 *addr) | |||
|  * is_zero_ether_addr - Determine if give Ethernet address is all zeros. | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if the address is all zeroes. | ||||
|  * Return: true if the address is all zeroes. | ||||
|  * | ||||
|  * Please note: addr must be aligned to u16. | ||||
|  */ | ||||
|  | @ -123,7 +123,7 @@ static inline bool is_zero_ether_addr(const u8 *addr) | |||
|  * is_multicast_ether_addr - Determine if the Ethernet address is a multicast. | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if the address is a multicast address. | ||||
|  * Return: true if the address is a multicast address. | ||||
|  * By definition the broadcast address is also a multicast address. | ||||
|  */ | ||||
| static inline bool is_multicast_ether_addr(const u8 *addr) | ||||
|  | @ -157,7 +157,7 @@ static inline bool is_multicast_ether_addr_64bits(const u8 *addr) | |||
|  * is_local_ether_addr - Determine if the Ethernet address is locally-assigned one (IEEE 802). | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if the address is a local address. | ||||
|  * Return: true if the address is a local address. | ||||
|  */ | ||||
| static inline bool is_local_ether_addr(const u8 *addr) | ||||
| { | ||||
|  | @ -168,7 +168,7 @@ static inline bool is_local_ether_addr(const u8 *addr) | |||
|  * is_broadcast_ether_addr - Determine if the Ethernet address is broadcast | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if the address is the broadcast address. | ||||
|  * Return: true if the address is the broadcast address. | ||||
|  * | ||||
|  * Please note: addr must be aligned to u16. | ||||
|  */ | ||||
|  | @ -183,7 +183,7 @@ static inline bool is_broadcast_ether_addr(const u8 *addr) | |||
|  * is_unicast_ether_addr - Determine if the Ethernet address is unicast | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return true if the address is a unicast address. | ||||
|  * Return: true if the address is a unicast address. | ||||
|  */ | ||||
| static inline bool is_unicast_ether_addr(const u8 *addr) | ||||
| { | ||||
|  | @ -197,7 +197,7 @@ static inline bool is_unicast_ether_addr(const u8 *addr) | |||
|  * Check that the Ethernet address (MAC) is not 00:00:00:00:00:00, is not | ||||
|  * a multicast address, and is not FF:FF:FF:FF:FF:FF. | ||||
|  * | ||||
|  * Return true if the address is valid. | ||||
|  * Return: true if the address is valid. | ||||
|  * | ||||
|  * Please note: addr must be aligned to u16. | ||||
|  */ | ||||
|  | @ -214,7 +214,7 @@ static inline bool is_valid_ether_addr(const u8 *addr) | |||
|  * | ||||
|  * Check that the value from the Ethertype/length field is a valid Ethertype. | ||||
|  * | ||||
|  * Return true if the valid is an 802.3 supported Ethertype. | ||||
|  * Return: true if the valid is an 802.3 supported Ethertype. | ||||
|  */ | ||||
| static inline bool eth_proto_is_802_3(__be16 proto) | ||||
| { | ||||
|  | @ -458,7 +458,7 @@ static inline bool ether_addr_is_ip_mcast(const u8 *addr) | |||
|  * ether_addr_to_u64 - Convert an Ethernet address into a u64 value. | ||||
|  * @addr: Pointer to a six-byte array containing the Ethernet address | ||||
|  * | ||||
|  * Return a u64 value of the address | ||||
|  * Return: a u64 value of the address | ||||
|  */ | ||||
| static inline u64 ether_addr_to_u64(const u8 *addr) | ||||
| { | ||||
|  |  | |||
|  | @ -257,7 +257,7 @@ struct ethtool_link_ksettings { | |||
|  *   @mode : one of the ETHTOOL_LINK_MODE_*_BIT | ||||
|  * (not atomic, no bound checking) | ||||
|  * | ||||
|  * Returns true/false. | ||||
|  * Returns: true/false. | ||||
|  */ | ||||
| #define ethtool_link_ksettings_test_link_mode(ptr, name, mode)		\ | ||||
| 	test_bit(ETHTOOL_LINK_MODE_ ## mode ## _BIT, (ptr)->link_modes.name) | ||||
|  | @ -1199,7 +1199,7 @@ ethtool_params_from_link_mode(struct ethtool_link_ksettings *link_ksettings, | |||
|  * @dev: pointer to net_device structure | ||||
|  * @vclock_index: pointer to pointer of vclock index | ||||
|  * | ||||
|  * Return number of phc vclocks | ||||
|  * Return: number of phc vclocks | ||||
|  */ | ||||
| int ethtool_get_phc_vclocks(struct net_device *dev, int **vclock_index); | ||||
| 
 | ||||
|  | @ -1253,7 +1253,7 @@ static inline int ethtool_mm_frag_size_min_to_add(u32 val_min, u32 *val_add, | |||
|  * ethtool_get_ts_info_by_layer - Obtains time stamping capabilities from the MAC or PHY layer. | ||||
|  * @dev: pointer to net_device structure | ||||
|  * @info: buffer to hold the result | ||||
|  * Returns zero on success, non-zero otherwise. | ||||
|  * Returns: zero on success, non-zero otherwise. | ||||
|  */ | ||||
| int ethtool_get_ts_info_by_layer(struct net_device *dev, | ||||
| 				 struct kernel_ethtool_ts_info *info); | ||||
|  |  | |||
|  | @ -310,7 +310,7 @@ static inline bool vlan_uses_dev(const struct net_device *dev) | |||
|  * eth_type_vlan - check for valid vlan ether type. | ||||
|  * @ethertype: ether type to check | ||||
|  * | ||||
|  * Returns true if the ether type is a vlan ether type. | ||||
|  * Returns: true if the ether type is a vlan ether type. | ||||
|  */ | ||||
| static inline bool eth_type_vlan(__be16 ethertype) | ||||
| { | ||||
|  | @ -341,9 +341,9 @@ static inline bool vlan_hw_offload_capable(netdev_features_t features, | |||
|  * @mac_len: MAC header length including outer vlan headers | ||||
|  * | ||||
|  * Inserts the VLAN tag into @skb as part of the payload at offset mac_len | ||||
|  * Returns error if skb_cow_head fails. | ||||
|  * | ||||
|  * Does not change skb->protocol so this function can be used during receive. | ||||
|  * | ||||
|  * Returns: error if skb_cow_head fails. | ||||
|  */ | ||||
| static inline int __vlan_insert_inner_tag(struct sk_buff *skb, | ||||
| 					  __be16 vlan_proto, u16 vlan_tci, | ||||
|  | @ -390,9 +390,9 @@ static inline int __vlan_insert_inner_tag(struct sk_buff *skb, | |||
|  * @vlan_tci: VLAN TCI to insert | ||||
|  * | ||||
|  * Inserts the VLAN tag into @skb as part of the payload | ||||
|  * Returns error if skb_cow_head fails. | ||||
|  * | ||||
|  * Does not change skb->protocol so this function can be used during receive. | ||||
|  * | ||||
|  * Returns: error if skb_cow_head fails. | ||||
|  */ | ||||
| static inline int __vlan_insert_tag(struct sk_buff *skb, | ||||
| 				    __be16 vlan_proto, u16 vlan_tci) | ||||
|  | @ -533,7 +533,7 @@ static inline void __vlan_hwaccel_put_tag(struct sk_buff *skb, | |||
|  * @skb: skbuff to query | ||||
|  * @vlan_tci: buffer to store value | ||||
|  * | ||||
|  * Returns error if the skb is not of VLAN type | ||||
|  * Returns: error if the skb is not of VLAN type | ||||
|  */ | ||||
| static inline int __vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) | ||||
| { | ||||
|  | @ -551,7 +551,7 @@ static inline int __vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) | |||
|  * @skb: skbuff to query | ||||
|  * @vlan_tci: buffer to store value | ||||
|  * | ||||
|  * Returns error if @skb->vlan_tci is not set correctly | ||||
|  * Returns: error if @skb->vlan_tci is not set correctly | ||||
|  */ | ||||
| static inline int __vlan_hwaccel_get_tag(const struct sk_buff *skb, | ||||
| 					 u16 *vlan_tci) | ||||
|  | @ -570,7 +570,7 @@ static inline int __vlan_hwaccel_get_tag(const struct sk_buff *skb, | |||
|  * @skb: skbuff to query | ||||
|  * @vlan_tci: buffer to store value | ||||
|  * | ||||
|  * Returns error if the skb is not VLAN tagged | ||||
|  * Returns: error if the skb is not VLAN tagged | ||||
|  */ | ||||
| static inline int vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) | ||||
| { | ||||
|  | @ -587,7 +587,7 @@ static inline int vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) | |||
|  * @type: first vlan protocol | ||||
|  * @depth: buffer to store length of eth and vlan tags in bytes | ||||
|  * | ||||
|  * Returns the EtherType of the packet, regardless of whether it is | ||||
|  * Returns: the EtherType of the packet, regardless of whether it is | ||||
|  * vlan encapsulated (normal or hardware accelerated) or not. | ||||
|  */ | ||||
| static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type, | ||||
|  | @ -629,7 +629,7 @@ static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type, | |||
|  * vlan_get_protocol - get protocol EtherType. | ||||
|  * @skb: skbuff to query | ||||
|  * | ||||
|  * Returns the EtherType of the packet, regardless of whether it is | ||||
|  * Returns: the EtherType of the packet, regardless of whether it is | ||||
|  * vlan encapsulated (normal or hardware accelerated) or not. | ||||
|  */ | ||||
| static inline __be16 vlan_get_protocol(const struct sk_buff *skb) | ||||
|  | @ -710,7 +710,7 @@ static inline void vlan_set_encap_proto(struct sk_buff *skb, | |||
|  * Expects the skb to contain a VLAN tag in the payload, and to have skb->data | ||||
|  * pointing at the MAC header. | ||||
|  * | ||||
|  * Returns a new pointer to skb->data, or NULL on failure to pull. | ||||
|  * Returns: a new pointer to skb->data, or NULL on failure to pull. | ||||
|  */ | ||||
| static inline void *vlan_remove_tag(struct sk_buff *skb, u16 *vlan_tci) | ||||
| { | ||||
|  | @ -727,7 +727,7 @@ static inline void *vlan_remove_tag(struct sk_buff *skb, u16 *vlan_tci) | |||
|  * skb_vlan_tagged - check if skb is vlan tagged. | ||||
|  * @skb: skbuff to query | ||||
|  * | ||||
|  * Returns true if the skb is tagged, regardless of whether it is hardware | ||||
|  * Returns: true if the skb is tagged, regardless of whether it is hardware | ||||
|  * accelerated or not. | ||||
|  */ | ||||
| static inline bool skb_vlan_tagged(const struct sk_buff *skb) | ||||
|  | @ -743,7 +743,7 @@ static inline bool skb_vlan_tagged(const struct sk_buff *skb) | |||
|  * skb_vlan_tagged_multi - check if skb is vlan tagged with multiple headers. | ||||
|  * @skb: skbuff to query | ||||
|  * | ||||
|  * Returns true if the skb is tagged with multiple vlan headers, regardless | ||||
|  * Returns: true if the skb is tagged with multiple vlan headers, regardless | ||||
|  * of whether it is hardware accelerated or not. | ||||
|  */ | ||||
| static inline bool skb_vlan_tagged_multi(struct sk_buff *skb) | ||||
|  | @ -774,7 +774,7 @@ static inline bool skb_vlan_tagged_multi(struct sk_buff *skb) | |||
|  * @skb: skbuff to query | ||||
|  * @features: features to be checked | ||||
|  * | ||||
|  * Returns features without unsafe ones if the skb has multiple tags. | ||||
|  * Returns: features without unsafe ones if the skb has multiple tags. | ||||
|  */ | ||||
| static inline netdev_features_t vlan_features_check(struct sk_buff *skb, | ||||
| 						    netdev_features_t features) | ||||
|  |  | |||
|  | @ -509,7 +509,7 @@ static inline bool napi_prefer_busy_poll(struct napi_struct *n) | |||
|  * is scheduled for example in the context of delayed timer | ||||
|  * that can be skipped if a NAPI is already scheduled. | ||||
|  * | ||||
|  * Return True if NAPI is scheduled, False otherwise. | ||||
|  * Return: True if NAPI is scheduled, False otherwise. | ||||
|  */ | ||||
| static inline bool napi_is_scheduled(struct napi_struct *n) | ||||
| { | ||||
|  | @ -524,7 +524,7 @@ bool napi_schedule_prep(struct napi_struct *n); | |||
|  * | ||||
|  * Schedule NAPI poll routine to be called if it is not already | ||||
|  * running. | ||||
|  * Return true if we schedule a NAPI or false if not. | ||||
|  * Return: true if we schedule a NAPI or false if not. | ||||
|  * Refer to napi_schedule_prep() for additional reason on why | ||||
|  * a NAPI might not be scheduled. | ||||
|  */ | ||||
|  | @ -558,7 +558,7 @@ static inline void napi_schedule_irqoff(struct napi_struct *n) | |||
|  * Mark NAPI processing as complete. Should only be called if poll budget | ||||
|  * has not been completely consumed. | ||||
|  * Prefer over napi_complete(). | ||||
|  * Return false if device should avoid rearming interrupts. | ||||
|  * Return: false if device should avoid rearming interrupts. | ||||
|  */ | ||||
| bool napi_complete_done(struct napi_struct *n, int work_done); | ||||
| 
 | ||||
|  | @ -3851,7 +3851,7 @@ static inline bool netif_attr_test_mask(unsigned long j, | |||
|  *	@online_mask: bitmask for CPUs/Rx queues that are online | ||||
|  *	@nr_bits: number of bits in the bitmask | ||||
|  * | ||||
|  * Returns true if a CPU/Rx queue is online. | ||||
|  * Returns: true if a CPU/Rx queue is online. | ||||
|  */ | ||||
| static inline bool netif_attr_test_online(unsigned long j, | ||||
| 					  const unsigned long *online_mask, | ||||
|  | @ -3871,7 +3871,8 @@ static inline bool netif_attr_test_online(unsigned long j, | |||
|  *	@srcp: the cpumask/Rx queue mask pointer | ||||
|  *	@nr_bits: number of bits in the bitmask | ||||
|  * | ||||
|  * Returns >= nr_bits if no further CPUs/Rx queues set. | ||||
|  * Returns: next (after n) CPU/Rx queue index in the mask; | ||||
|  * >= nr_bits if no further CPUs/Rx queues set. | ||||
|  */ | ||||
| static inline unsigned int netif_attrmask_next(int n, const unsigned long *srcp, | ||||
| 					       unsigned int nr_bits) | ||||
|  | @ -3893,7 +3894,8 @@ static inline unsigned int netif_attrmask_next(int n, const unsigned long *srcp, | |||
|  *	@src2p: the second CPUs/Rx queues mask pointer | ||||
|  *	@nr_bits: number of bits in the bitmask | ||||
|  * | ||||
|  * Returns >= nr_bits if no further CPUs/Rx queues set in both. | ||||
|  * Returns: next (after n) CPU/Rx queue index set in both masks; | ||||
|  * >= nr_bits if no further CPUs/Rx queues set in both. | ||||
|  */ | ||||
| static inline int netif_attrmask_next_and(int n, const unsigned long *src1p, | ||||
| 					  const unsigned long *src2p, | ||||
|  |  | |||
|  | @ -357,7 +357,7 @@ extern struct static_key xt_tee_enabled; | |||
|  * Begin packet processing : all readers must wait the end | ||||
|  * 1) Must be called with preemption disabled | ||||
|  * 2) softirqs must be disabled too (or we should use this_cpu_add()) | ||||
|  * Returns : | ||||
|  * Returns: | ||||
|  *  1 if no recursion on this cpu | ||||
|  *  0 if recursion detected | ||||
|  */ | ||||
|  |  | |||
|  | @ -66,7 +66,6 @@ static inline bool nf_hook_egress_active(void) | |||
|  * @rc: result code which shall be returned by __dev_queue_xmit() on failure | ||||
|  * @dev: netdev whose egress hooks shall be applied to @skb | ||||
|  * | ||||
|  * Returns @skb on success or %NULL if the packet was consumed or filtered. | ||||
|  * Caller must hold rcu_read_lock. | ||||
|  * | ||||
|  * On ingress, packets are classified first by tc, then by netfilter. | ||||
|  | @ -81,6 +80,8 @@ static inline bool nf_hook_egress_active(void) | |||
|  * called recursively by tunnel drivers such as vxlan, the flag is reverted to | ||||
|  * false after sch_handle_egress().  This ensures that netfilter is applied | ||||
|  * both on the overlay and underlying network. | ||||
|  * | ||||
|  * Returns: @skb on success or %NULL if the packet was consumed or filtered. | ||||
|  */ | ||||
| static inline struct sk_buff *nf_hook_egress(struct sk_buff *skb, int *rc, | ||||
| 					     struct net_device *dev) | ||||
|  |  | |||
|  | @ -307,7 +307,7 @@ static inline u64 adjust_by_scaled_ppm(u64 base, long scaled_ppm) | |||
|  * @info:   Structure describing the new clock. | ||||
|  * @parent: Pointer to the parent device of the new clock. | ||||
|  * | ||||
|  * Returns a valid pointer on success or PTR_ERR on failure.  If PHC | ||||
|  * Returns: a valid pointer on success or PTR_ERR on failure.  If PHC | ||||
|  * support is missing at the configuration level, this function | ||||
|  * returns NULL, and drivers are expected to gracefully handle that | ||||
|  * case separately. | ||||
|  | @ -445,7 +445,7 @@ int ptp_get_vclocks_index(int pclock_index, int **vclock_index); | |||
|  * @hwtstamp:     timestamp | ||||
|  * @vclock_index: phc index of ptp vclock. | ||||
|  * | ||||
|  * Returns converted timestamp, or 0 on error. | ||||
|  * Returns: converted timestamp, or 0 on error. | ||||
|  */ | ||||
| ktime_t ptp_convert_timestamp(const ktime_t *hwtstamp, int vclock_index); | ||||
| #else | ||||
|  |  | |||
|  | @ -241,7 +241,7 @@ bool rfkill_soft_blocked(struct rfkill *rfkill); | |||
|  * rfkill_find_type - Helper for finding rfkill type by name | ||||
|  * @name: the name of the type | ||||
|  * | ||||
|  * Returns enum rfkill_type that corresponds to the name. | ||||
|  * Returns: enum rfkill_type that corresponds to the name. | ||||
|  */ | ||||
| enum rfkill_type rfkill_find_type(const char *name); | ||||
| 
 | ||||
|  |  | |||
|  | @ -78,7 +78,7 @@ static inline bool lockdep_rtnl_is_held(void) | |||
|  * rtnl_dereference - fetch RCU pointer when updates are prevented by RTNL | ||||
|  * @p: The pointer to read, prior to dereferencing | ||||
|  * | ||||
|  * Return the value of the specified RCU-protected pointer, but omit | ||||
|  * Return: the value of the specified RCU-protected pointer, but omit | ||||
|  * the READ_ONCE(), because caller holds RTNL. | ||||
|  */ | ||||
| #define rtnl_dereference(p)					\ | ||||
|  |  | |||
|  | @ -1134,7 +1134,7 @@ static inline bool skb_pfmemalloc(const struct sk_buff *skb) | |||
|  * skb_dst - returns skb dst_entry | ||||
|  * @skb: buffer | ||||
|  * | ||||
|  * Returns skb dst_entry, regardless of reference taken or not. | ||||
|  * Returns: skb dst_entry, regardless of reference taken or not. | ||||
|  */ | ||||
| static inline struct dst_entry *skb_dst(const struct sk_buff *skb) | ||||
| { | ||||
|  | @ -1222,7 +1222,7 @@ static inline bool skb_wifi_acked_valid(const struct sk_buff *skb) | |||
|  * skb_unref - decrement the skb's reference count | ||||
|  * @skb: buffer | ||||
|  * | ||||
|  * Returns true if we can free the skb. | ||||
|  * Returns: true if we can free the skb. | ||||
|  */ | ||||
| static inline bool skb_unref(struct sk_buff *skb) | ||||
| { | ||||
|  | @ -1344,7 +1344,7 @@ struct sk_buff_fclones { | |||
|  *	@sk: socket | ||||
|  *	@skb: buffer | ||||
|  * | ||||
|  * Returns true if skb is a fast clone, and its clone is not freed. | ||||
|  * Returns: true if skb is a fast clone, and its clone is not freed. | ||||
|  * Some drivers call skb_orphan() in their ndo_start_xmit(), | ||||
|  * so we also check that didn't happen. | ||||
|  */ | ||||
|  | @ -3516,7 +3516,7 @@ static inline struct page *__dev_alloc_page_noprof(gfp_t gfp_mask) | |||
|  * A page shouldn't be considered for reusing/recycling if it was allocated | ||||
|  * under memory pressure or at a distant memory node. | ||||
|  * | ||||
|  * Returns false if this page should be returned to page allocator, true | ||||
|  * Returns: false if this page should be returned to page allocator, true | ||||
|  * otherwise. | ||||
|  */ | ||||
| static inline bool dev_page_is_reusable(const struct page *page) | ||||
|  | @ -3633,7 +3633,7 @@ int skb_cow_data_for_xdp(struct page_pool *pool, struct sk_buff **pskb, | |||
|  * skb_frag_address - gets the address of the data contained in a paged fragment | ||||
|  * @frag: the paged fragment buffer | ||||
|  * | ||||
|  * Returns the address of the data within @frag. The page must already | ||||
|  * Returns: the address of the data within @frag. The page must already | ||||
|  * be mapped. | ||||
|  */ | ||||
| static inline void *skb_frag_address(const skb_frag_t *frag) | ||||
|  | @ -3648,7 +3648,7 @@ static inline void *skb_frag_address(const skb_frag_t *frag) | |||
|  * skb_frag_address_safe - gets the address of the data contained in a paged fragment | ||||
|  * @frag: the paged fragment buffer | ||||
|  * | ||||
|  * Returns the address of the data within @frag. Checks that the page | ||||
|  * Returns: the address of the data within @frag. Checks that the page | ||||
|  * is mapped and returns %NULL otherwise. | ||||
|  */ | ||||
| static inline void *skb_frag_address_safe(const skb_frag_t *frag) | ||||
|  | @ -3890,7 +3890,7 @@ static inline int skb_linearize(struct sk_buff *skb) | |||
|  * skb_has_shared_frag - can any frag be overwritten | ||||
|  * @skb: buffer to test | ||||
|  * | ||||
|  * Return true if the skb has at least one frag that might be modified | ||||
|  * Return: true if the skb has at least one frag that might be modified | ||||
|  * by an external entity (as in vmsplice()/sendfile()) | ||||
|  */ | ||||
| static inline bool skb_has_shared_frag(const struct sk_buff *skb) | ||||
|  | @ -4612,7 +4612,7 @@ static inline void __skb_reset_checksum_unnecessary(struct sk_buff *skb) | |||
| 
 | ||||
| /* Check if we need to perform checksum complete validation.
 | ||||
|  * | ||||
|  * Returns true if checksum complete is needed, false otherwise | ||||
|  * Returns: true if checksum complete is needed, false otherwise | ||||
|  * (either checksum is unnecessary or zero checksum is allowed). | ||||
|  */ | ||||
| static inline bool __skb_checksum_validate_needed(struct sk_buff *skb, | ||||
|  |  | |||
|  | @ -97,7 +97,7 @@ struct wwan_port_caps { | |||
|  * | ||||
|  * This function must be balanced with a call to wwan_remove_port(). | ||||
|  * | ||||
|  * Returns a valid pointer to wwan_port on success or PTR_ERR on failure | ||||
|  * Returns: a valid pointer to wwan_port on success or PTR_ERR on failure | ||||
|  */ | ||||
| struct wwan_port *wwan_create_port(struct device *parent, | ||||
| 				   enum wwan_port_type type, | ||||
|  |  | |||
|  | @ -5957,7 +5957,7 @@ int wiphy_register(struct wiphy *wiphy); | |||
|  * @wiphy: the wiphy to check the locking on | ||||
|  * @p: The pointer to read, prior to dereferencing | ||||
|  * | ||||
|  * Return the value of the specified RCU-protected pointer, but omit the | ||||
|  * Return: the value of the specified RCU-protected pointer, but omit the | ||||
|  * READ_ONCE(), because caller holds the wiphy mutex used for updates. | ||||
|  */ | ||||
| #define wiphy_dereference(wiphy, p)				\ | ||||
|  |  | |||
|  | @ -307,7 +307,7 @@ static inline bool dst_hold_safe(struct dst_entry *dst) | |||
|  * @skb: buffer | ||||
|  * | ||||
|  * If dst is not yet refcounted and not destroyed, grab a ref on it. | ||||
|  * Returns true if dst is refcounted. | ||||
|  * Returns: true if dst is refcounted. | ||||
|  */ | ||||
| static inline bool skb_dst_force(struct sk_buff *skb) | ||||
| { | ||||
|  |  | |||
|  | @ -354,7 +354,7 @@ __genlmsg_iput(struct sk_buff *skb, const struct genl_info *info, int flags) | |||
|  * such requests) or a struct initialized by genl_info_init_ntf() | ||||
|  * when constructing notifications. | ||||
|  * | ||||
|  * Returns pointer to new genetlink header. | ||||
|  * Returns: pointer to new genetlink header. | ||||
|  */ | ||||
| static inline void * | ||||
| genlmsg_iput(struct sk_buff *skb, const struct genl_info *info) | ||||
|  | @ -366,7 +366,7 @@ genlmsg_iput(struct sk_buff *skb, const struct genl_info *info) | |||
|  * genlmsg_nlhdr - Obtain netlink header from user specified header | ||||
|  * @user_hdr: user header as returned from genlmsg_put() | ||||
|  * | ||||
|  * Returns pointer to netlink header. | ||||
|  * Returns: pointer to netlink header. | ||||
|  */ | ||||
| static inline struct nlmsghdr *genlmsg_nlhdr(void *user_hdr) | ||||
| { | ||||
|  | @ -435,7 +435,7 @@ static inline void genl_dump_check_consistent(struct netlink_callback *cb, | |||
|  * @flags: netlink message flags | ||||
|  * @cmd: generic netlink command | ||||
|  * | ||||
|  * Returns pointer to user specific header | ||||
|  * Returns: pointer to user specific header | ||||
|  */ | ||||
| static inline void *genlmsg_put_reply(struct sk_buff *skb, | ||||
| 				      struct genl_info *info, | ||||
|  |  | |||
|  | @ -471,7 +471,7 @@ struct ipv6_txoptions *ipv6_update_options(struct sock *sk, | |||
| /* This helper is specialized for BIG TCP needs.
 | ||||
|  * It assumes the hop_jumbo_hdr will immediately follow the IPV6 header. | ||||
|  * It assumes headers are already in skb->head. | ||||
|  * Returns 0, or IPPROTO_TCP if a BIG TCP packet is there. | ||||
|  * Returns: 0, or IPPROTO_TCP if a BIG TCP packet is there. | ||||
|  */ | ||||
| static inline int ipv6_has_hopopt_jumbo(const struct sk_buff *skb) | ||||
| { | ||||
|  |  | |||
|  | @ -202,7 +202,7 @@ struct iucv_handler { | |||
|  * | ||||
|  * Registers a driver with IUCV. | ||||
|  * | ||||
|  * Returns 0 on success, -ENOMEM if the memory allocation for the pathid | ||||
|  * Returns: 0 on success, -ENOMEM if the memory allocation for the pathid | ||||
|  * table failed, or -EIO if IUCV_DECLARE_BUFFER failed on all cpus. | ||||
|  */ | ||||
| int iucv_register(struct iucv_handler *handler, int smp); | ||||
|  | @ -224,7 +224,7 @@ void iucv_unregister(struct iucv_handler *handle, int smp); | |||
|  * | ||||
|  * Allocate a new path structure for use with iucv_connect. | ||||
|  * | ||||
|  * Returns NULL if the memory allocation failed or a pointer to the | ||||
|  * Returns: NULL if the memory allocation failed or a pointer to the | ||||
|  * path structure. | ||||
|  */ | ||||
| static inline struct iucv_path *iucv_path_alloc(u16 msglim, u8 flags, gfp_t gfp) | ||||
|  | @ -260,7 +260,7 @@ static inline void iucv_path_free(struct iucv_path *path) | |||
|  * This function is issued after the user received a connection pending | ||||
|  * external interrupt and now wishes to complete the IUCV communication path. | ||||
|  * | ||||
|  * Returns the result of the CP IUCV call. | ||||
|  * Returns: the result of the CP IUCV call. | ||||
|  */ | ||||
| int iucv_path_accept(struct iucv_path *path, struct iucv_handler *handler, | ||||
| 		     u8 *userdata, void *private); | ||||
|  | @ -278,7 +278,7 @@ int iucv_path_accept(struct iucv_path *path, struct iucv_handler *handler, | |||
|  * successfully, you are not able to use the path until you receive an IUCV | ||||
|  * Connection Complete external interrupt. | ||||
|  * | ||||
|  * Returns the result of the CP IUCV call. | ||||
|  * Returns: the result of the CP IUCV call. | ||||
|  */ | ||||
| int iucv_path_connect(struct iucv_path *path, struct iucv_handler *handler, | ||||
| 		      u8 *userid, u8 *system, u8 *userdata, | ||||
|  | @ -292,7 +292,7 @@ int iucv_path_connect(struct iucv_path *path, struct iucv_handler *handler, | |||
|  * This function temporarily suspends incoming messages on an IUCV path. | ||||
|  * You can later reactivate the path by invoking the iucv_resume function. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_path_quiesce(struct iucv_path *path, u8 *userdata); | ||||
| 
 | ||||
|  | @ -304,7 +304,7 @@ int iucv_path_quiesce(struct iucv_path *path, u8 *userdata); | |||
|  * This function resumes incoming messages on an IUCV path that has | ||||
|  * been stopped with iucv_path_quiesce. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_path_resume(struct iucv_path *path, u8 *userdata); | ||||
| 
 | ||||
|  | @ -315,7 +315,7 @@ int iucv_path_resume(struct iucv_path *path, u8 *userdata); | |||
|  * | ||||
|  * This function terminates an IUCV path. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_path_sever(struct iucv_path *path, u8 *userdata); | ||||
| 
 | ||||
|  | @ -327,7 +327,7 @@ int iucv_path_sever(struct iucv_path *path, u8 *userdata); | |||
|  * | ||||
|  * Cancels a message you have sent. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_purge(struct iucv_path *path, struct iucv_message *msg, | ||||
| 		       u32 srccls); | ||||
|  | @ -347,7 +347,7 @@ int iucv_message_purge(struct iucv_path *path, struct iucv_message *msg, | |||
|  * | ||||
|  * Locking:	local_bh_enable/local_bh_disable | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_receive(struct iucv_path *path, struct iucv_message *msg, | ||||
| 			 u8 flags, void *buffer, size_t size, size_t *residual); | ||||
|  | @ -367,7 +367,7 @@ int iucv_message_receive(struct iucv_path *path, struct iucv_message *msg, | |||
|  * | ||||
|  * Locking:	no locking. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int __iucv_message_receive(struct iucv_path *path, struct iucv_message *msg, | ||||
| 			   u8 flags, void *buffer, size_t size, | ||||
|  | @ -382,7 +382,7 @@ int __iucv_message_receive(struct iucv_path *path, struct iucv_message *msg, | |||
|  * are notified of a message and the time that you complete the message, | ||||
|  * the message may be rejected. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_reject(struct iucv_path *path, struct iucv_message *msg); | ||||
| 
 | ||||
|  | @ -399,7 +399,7 @@ int iucv_message_reject(struct iucv_path *path, struct iucv_message *msg); | |||
|  * pathid, msgid, and trgcls. Prmmsg signifies the data is moved into | ||||
|  * the parameter list. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_reply(struct iucv_path *path, struct iucv_message *msg, | ||||
| 		       u8 flags, void *reply, size_t size); | ||||
|  | @ -419,7 +419,7 @@ int iucv_message_reply(struct iucv_path *path, struct iucv_message *msg, | |||
|  * | ||||
|  * Locking:	local_bh_enable/local_bh_disable | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_send(struct iucv_path *path, struct iucv_message *msg, | ||||
| 		      u8 flags, u32 srccls, void *buffer, size_t size); | ||||
|  | @ -439,7 +439,7 @@ int iucv_message_send(struct iucv_path *path, struct iucv_message *msg, | |||
|  * | ||||
|  * Locking:	no locking. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int __iucv_message_send(struct iucv_path *path, struct iucv_message *msg, | ||||
| 			u8 flags, u32 srccls, void *buffer, size_t size); | ||||
|  | @ -461,7 +461,7 @@ int __iucv_message_send(struct iucv_path *path, struct iucv_message *msg, | |||
|  * reply to the message and a buffer is provided into which IUCV moves | ||||
|  * the reply to this message. | ||||
|  * | ||||
|  * Returns the result from the CP IUCV call. | ||||
|  * Returns: the result from the CP IUCV call. | ||||
|  */ | ||||
| int iucv_message_send2way(struct iucv_path *path, struct iucv_message *msg, | ||||
| 			  u8 flags, u32 srccls, void *buffer, size_t size, | ||||
|  |  | |||
|  | @ -49,7 +49,7 @@ __be32 nf_tproxy_laddr4(struct sk_buff *skb, __be32 user_laddr, __be32 daddr); | |||
|  * | ||||
|  * nf_tproxy_handle_time_wait4() consumes the socket reference passed in. | ||||
|  * | ||||
|  * Returns the listener socket if there's one, the TIME_WAIT socket if | ||||
|  * Returns: the listener socket if there's one, the TIME_WAIT socket if | ||||
|  * no such listener is found, or NULL if the TCP header is incomplete. | ||||
|  */ | ||||
| struct sock * | ||||
|  | @ -108,7 +108,7 @@ nf_tproxy_laddr6(struct sk_buff *skb, const struct in6_addr *user_laddr, | |||
|  * | ||||
|  * nf_tproxy_handle_time_wait6() consumes the socket reference passed in. | ||||
|  * | ||||
|  * Returns the listener socket if there's one, the TIME_WAIT socket if | ||||
|  * Returns: the listener socket if there's one, the TIME_WAIT socket if | ||||
|  * no such listener is found, or NULL if the TCP header is incomplete. | ||||
|  */ | ||||
| struct sock * | ||||
|  |  | |||
|  | @ -649,7 +649,7 @@ static inline int nlmsg_ok(const struct nlmsghdr *nlh, int remaining) | |||
|  * @nlh: netlink message header | ||||
|  * @remaining: number of bytes remaining in message stream | ||||
|  * | ||||
|  * Returns the next netlink message in the message stream and | ||||
|  * Returns: the next netlink message in the message stream and | ||||
|  * decrements remaining by the size of the current message. | ||||
|  */ | ||||
| static inline struct nlmsghdr * | ||||
|  | @ -676,7 +676,7 @@ nlmsg_next(const struct nlmsghdr *nlh, int *remaining) | |||
|  * exceeding maxtype will be rejected, policy must be specified, attributes | ||||
|  * will be validated in the strictest way possible. | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_parse(struct nlattr **tb, int maxtype, | ||||
| 			    const struct nlattr *head, int len, | ||||
|  | @ -701,7 +701,7 @@ static inline int nla_parse(struct nlattr **tb, int maxtype, | |||
|  * exceeding maxtype will be ignored and attributes from the policy are not | ||||
|  * always strictly validated (only for new attributes). | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_parse_deprecated(struct nlattr **tb, int maxtype, | ||||
| 				       const struct nlattr *head, int len, | ||||
|  | @ -726,7 +726,7 @@ static inline int nla_parse_deprecated(struct nlattr **tb, int maxtype, | |||
|  * exceeding maxtype will be rejected as well as trailing data, but the | ||||
|  * policy is not completely strictly validated (only for new attributes). | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_parse_deprecated_strict(struct nlattr **tb, int maxtype, | ||||
| 					      const struct nlattr *head, | ||||
|  | @ -833,7 +833,7 @@ nlmsg_parse_deprecated_strict(const struct nlmsghdr *nlh, int hdrlen, | |||
|  * @hdrlen: length of family specific header | ||||
|  * @attrtype: type of attribute to look for | ||||
|  * | ||||
|  * Returns the first attribute which matches the specified type. | ||||
|  * Returns: the first attribute which matches the specified type. | ||||
|  */ | ||||
| static inline struct nlattr *nlmsg_find_attr(const struct nlmsghdr *nlh, | ||||
| 					     int hdrlen, int attrtype) | ||||
|  | @ -854,7 +854,7 @@ static inline struct nlattr *nlmsg_find_attr(const struct nlmsghdr *nlh, | |||
|  * specified policy. Validation is done in liberal mode. | ||||
|  * See documentation of struct nla_policy for more details. | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_validate_deprecated(const struct nlattr *head, int len, | ||||
| 					  int maxtype, | ||||
|  | @ -877,7 +877,7 @@ static inline int nla_validate_deprecated(const struct nlattr *head, int len, | |||
|  * specified policy. Validation is done in strict mode. | ||||
|  * See documentation of struct nla_policy for more details. | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_validate(const struct nlattr *head, int len, int maxtype, | ||||
| 			       const struct nla_policy *policy, | ||||
|  | @ -914,7 +914,7 @@ static inline int nlmsg_validate_deprecated(const struct nlmsghdr *nlh, | |||
|  * nlmsg_report - need to report back to application? | ||||
|  * @nlh: netlink message header | ||||
|  * | ||||
|  * Returns 1 if a report back to the application is requested. | ||||
|  * Returns: 1 if a report back to the application is requested. | ||||
|  */ | ||||
| static inline int nlmsg_report(const struct nlmsghdr *nlh) | ||||
| { | ||||
|  | @ -925,7 +925,7 @@ static inline int nlmsg_report(const struct nlmsghdr *nlh) | |||
|  * nlmsg_seq - return the seq number of netlink message | ||||
|  * @nlh: netlink message header | ||||
|  * | ||||
|  * Returns 0 if netlink message is NULL | ||||
|  * Returns: 0 if netlink message is NULL | ||||
|  */ | ||||
| static inline u32 nlmsg_seq(const struct nlmsghdr *nlh) | ||||
| { | ||||
|  | @ -952,7 +952,7 @@ static inline u32 nlmsg_seq(const struct nlmsghdr *nlh) | |||
|  * @payload: length of message payload | ||||
|  * @flags: message flags | ||||
|  * | ||||
|  * Returns NULL if the tailroom of the skb is insufficient to store | ||||
|  * Returns: NULL if the tailroom of the skb is insufficient to store | ||||
|  * the message header and payload. | ||||
|  */ | ||||
| static inline struct nlmsghdr *nlmsg_put(struct sk_buff *skb, u32 portid, u32 seq, | ||||
|  | @ -971,7 +971,7 @@ static inline struct nlmsghdr *nlmsg_put(struct sk_buff *skb, u32 portid, u32 se | |||
|  * | ||||
|  * Append data to an existing nlmsg, used when constructing a message | ||||
|  * with multiple fixed-format headers (which is rare). | ||||
|  * Returns NULL if the tailroom of the skb is insufficient to store | ||||
|  * Returns: NULL if the tailroom of the skb is insufficient to store | ||||
|  * the extra payload. | ||||
|  */ | ||||
| static inline void *nlmsg_append(struct sk_buff *skb, u32 size) | ||||
|  | @ -993,7 +993,7 @@ static inline void *nlmsg_append(struct sk_buff *skb, u32 size) | |||
|  * @payload: length of message payload | ||||
|  * @flags: message flags | ||||
|  * | ||||
|  * Returns NULL if the tailroom of the skb is insufficient to store | ||||
|  * Returns: NULL if the tailroom of the skb is insufficient to store | ||||
|  * the message header and payload. | ||||
|  */ | ||||
| static inline struct nlmsghdr *nlmsg_put_answer(struct sk_buff *skb, | ||||
|  | @ -1050,7 +1050,7 @@ static inline void nlmsg_end(struct sk_buff *skb, struct nlmsghdr *nlh) | |||
|  * nlmsg_get_pos - return current position in netlink message | ||||
|  * @skb: socket buffer the message is stored in | ||||
|  * | ||||
|  * Returns a pointer to the current tail of the message. | ||||
|  * Returns: a pointer to the current tail of the message. | ||||
|  */ | ||||
| static inline void *nlmsg_get_pos(struct sk_buff *skb) | ||||
| { | ||||
|  | @ -1276,7 +1276,7 @@ static inline int nla_ok(const struct nlattr *nla, int remaining) | |||
|  * @nla: netlink attribute | ||||
|  * @remaining: number of bytes remaining in attribute stream | ||||
|  * | ||||
|  * Returns the next netlink attribute in the attribute stream and | ||||
|  * Returns: the next netlink attribute in the attribute stream and | ||||
|  * decrements remaining by the size of the current attribute. | ||||
|  */ | ||||
| static inline struct nlattr *nla_next(const struct nlattr *nla, int *remaining) | ||||
|  | @ -1292,7 +1292,7 @@ static inline struct nlattr *nla_next(const struct nlattr *nla, int *remaining) | |||
|  * @nla: attribute containing the nested attributes | ||||
|  * @attrtype: type of attribute to look for | ||||
|  * | ||||
|  * Returns the first attribute which matches the specified type. | ||||
|  * Returns: the first attribute which matches the specified type. | ||||
|  */ | ||||
| static inline struct nlattr * | ||||
| nla_find_nested(const struct nlattr *nla, int attrtype) | ||||
|  | @ -2091,7 +2091,7 @@ static inline int nla_get_flag(const struct nlattr *nla) | |||
|  * nla_get_msecs - return payload of msecs attribute | ||||
|  * @nla: msecs netlink attribute | ||||
|  * | ||||
|  * Returns the number of milliseconds in jiffies. | ||||
|  * Returns: the number of milliseconds in jiffies. | ||||
|  */ | ||||
| static inline unsigned long nla_get_msecs(const struct nlattr *nla) | ||||
| { | ||||
|  | @ -2183,7 +2183,7 @@ static inline void *nla_memdup_noprof(const struct nlattr *src, gfp_t gfp) | |||
|  * marked their nest attributes with NLA_F_NESTED flag. New APIs should use | ||||
|  * nla_nest_start() which sets the flag. | ||||
|  * | ||||
|  * Returns the container attribute or NULL on error | ||||
|  * Returns: the container attribute or NULL on error | ||||
|  */ | ||||
| static inline struct nlattr *nla_nest_start_noflag(struct sk_buff *skb, | ||||
| 						   int attrtype) | ||||
|  | @ -2204,7 +2204,7 @@ static inline struct nlattr *nla_nest_start_noflag(struct sk_buff *skb, | |||
|  * Unlike nla_nest_start_noflag(), mark the nest attribute with NLA_F_NESTED | ||||
|  * flag. This is the preferred function to use in new code. | ||||
|  * | ||||
|  * Returns the container attribute or NULL on error | ||||
|  * Returns: the container attribute or NULL on error | ||||
|  */ | ||||
| static inline struct nlattr *nla_nest_start(struct sk_buff *skb, int attrtype) | ||||
| { | ||||
|  | @ -2219,7 +2219,7 @@ static inline struct nlattr *nla_nest_start(struct sk_buff *skb, int attrtype) | |||
|  * Corrects the container attribute header to include the all | ||||
|  * appended attributes. | ||||
|  * | ||||
|  * Returns the total data length of the skb. | ||||
|  * Returns: the total data length of the skb. | ||||
|  */ | ||||
| static inline int nla_nest_end(struct sk_buff *skb, struct nlattr *start) | ||||
| { | ||||
|  | @ -2252,7 +2252,7 @@ static inline void nla_nest_cancel(struct sk_buff *skb, struct nlattr *start) | |||
|  * specified policy. Attributes with a type exceeding maxtype will be | ||||
|  * ignored. See documentation of struct nla_policy for more details. | ||||
|  * | ||||
|  * Returns 0 on success or a negative error code. | ||||
|  * Returns: 0 on success or a negative error code. | ||||
|  */ | ||||
| static inline int __nla_validate_nested(const struct nlattr *start, int maxtype, | ||||
| 					const struct nla_policy *policy, | ||||
|  | @ -2285,7 +2285,7 @@ nla_validate_nested_deprecated(const struct nlattr *start, int maxtype, | |||
|  * nla_need_padding_for_64bit - test 64-bit alignment of the next attribute | ||||
|  * @skb: socket buffer the message is stored in | ||||
|  * | ||||
|  * Return true if padding is needed to align the next attribute (nla_data()) to | ||||
|  * Return: true if padding is needed to align the next attribute (nla_data()) to | ||||
|  * a 64-bit aligned area. | ||||
|  */ | ||||
| static inline bool nla_need_padding_for_64bit(struct sk_buff *skb) | ||||
|  | @ -2312,7 +2312,7 @@ static inline bool nla_need_padding_for_64bit(struct sk_buff *skb) | |||
|  * This will only be done in architectures which do not have | ||||
|  * CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS defined. | ||||
|  * | ||||
|  * Returns zero on success or a negative error code. | ||||
|  * Returns: zero on success or a negative error code. | ||||
|  */ | ||||
| static inline int nla_align_64bit(struct sk_buff *skb, int padattr) | ||||
| { | ||||
|  |  | |||
|  | @ -104,8 +104,7 @@ static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) | |||
|  * | ||||
|  * Get a page fragment from the page allocator or page_pool caches. | ||||
|  * | ||||
|  * Return: | ||||
|  * Return allocated page fragment, otherwise return NULL. | ||||
|  * Return: allocated page fragment, otherwise return NULL. | ||||
|  */ | ||||
| static inline struct page *page_pool_dev_alloc_frag(struct page_pool *pool, | ||||
| 						    unsigned int *offset, | ||||
|  | @ -155,8 +154,7 @@ static inline struct page *page_pool_alloc(struct page_pool *pool, | |||
|  * depending on the requested size in order to allocate memory with least memory | ||||
|  * utilization and performance penalty. | ||||
|  * | ||||
|  * Return: | ||||
|  * Return allocated page or page fragment, otherwise return NULL. | ||||
|  * Return: allocated page or page fragment, otherwise return NULL. | ||||
|  */ | ||||
| static inline struct page *page_pool_dev_alloc(struct page_pool *pool, | ||||
| 					       unsigned int *offset, | ||||
|  | @ -190,8 +188,7 @@ static inline void *page_pool_alloc_va(struct page_pool *pool, | |||
|  * This is just a thin wrapper around the page_pool_alloc() API, and | ||||
|  * it returns va of the allocated page or page fragment. | ||||
|  * | ||||
|  * Return: | ||||
|  * Return the va for the allocated page or page fragment, otherwise return NULL. | ||||
|  * Return: the va for the allocated page or page fragment, otherwise return NULL. | ||||
|  */ | ||||
| static inline void *page_pool_dev_alloc_va(struct page_pool *pool, | ||||
| 					   unsigned int *size) | ||||
|  |  | |||
|  | @ -319,7 +319,7 @@ tcf_exts_hw_stats_update(const struct tcf_exts *exts, | |||
|  * tcf_exts_has_actions - check if at least one action is present | ||||
|  * @exts: tc filter extensions handle | ||||
|  * | ||||
|  * Returns true if at least one action is present. | ||||
|  * Returns: true if at least one action is present. | ||||
|  */ | ||||
| static inline bool tcf_exts_has_actions(struct tcf_exts *exts) | ||||
| { | ||||
|  | @ -501,7 +501,7 @@ int __tcf_em_tree_match(struct sk_buff *, struct tcf_ematch_tree *, | |||
|  * through all ematches respecting their logic relations returning | ||||
|  * as soon as the result is obvious. | ||||
|  * | ||||
|  * Returns 1 if the ematch tree as-one matches, no ematches are configured | ||||
|  * Returns: 1 if the ematch tree as-one matches, no ematches are configured | ||||
|  * or ematch is not enabled in the kernel, otherwise 0 is returned. | ||||
|  */ | ||||
| static inline int tcf_em_tree_match(struct sk_buff *skb, | ||||
|  |  | |||
|  | @ -1817,7 +1817,7 @@ int tcp_sigpool_hash_skb_data(struct tcp_sigpool *hp, | |||
|  * @id: tcp_sigpool that was previously allocated by tcp_sigpool_alloc_ahash() | ||||
|  * @c: returned tcp_sigpool for usage (uninitialized on failure) | ||||
|  * | ||||
|  * Returns 0 on success, error otherwise. | ||||
|  * Returns: 0 on success, error otherwise. | ||||
|  */ | ||||
| int tcp_sigpool_start(unsigned int id, struct tcp_sigpool *c); | ||||
| /**
 | ||||
|  |  | |||
		Loading…
	
		Reference in a new issue
	
	 Jakub Kicinski
						Jakub Kicinski