mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 06:01:57 +00:00
trivial: fix then -> than typos in comments and documentation
- (better, more, bigger ...) then -> (...) than Signed-off-by: Frederik Schwarzer <schwarzerf@gmail.com> Signed-off-by: Jiri Kosina <jkosina@suse.cz>
This commit is contained in:
parent
0abb8b6a93
commit
025dfdafe7
@ -74,7 +74,7 @@ a sensor.
|
||||
Notice that some banks have both a read and a write address this is how the
|
||||
uGuru determines if a read from or a write to the bank is taking place, thus
|
||||
when reading you should always use the read address and when writing the
|
||||
write address. The write address is always one (1) more then the read address.
|
||||
write address. The write address is always one (1) more than the read address.
|
||||
|
||||
|
||||
uGuru ready
|
||||
@ -224,7 +224,7 @@ Bit 3: Beep if alarm (RW)
|
||||
Bit 4: 1 if alarm cause measured temp is over the warning threshold (R)
|
||||
Bit 5: 1 if alarm cause measured volt is over the max threshold (R)
|
||||
Bit 6: 1 if alarm cause measured volt is under the min threshold (R)
|
||||
Bit 7: Volt sensor: Shutdown if alarm persist for more then 4 seconds (RW)
|
||||
Bit 7: Volt sensor: Shutdown if alarm persist for more than 4 seconds (RW)
|
||||
Temp sensor: Shutdown if temp is over the shutdown threshold (RW)
|
||||
|
||||
* This bit is only honored/used by the uGuru if a temp sensor is connected
|
||||
@ -293,7 +293,7 @@ Byte 0:
|
||||
Alarm behaviour for the selected sensor. A 1 enables the described behaviour.
|
||||
Bit 0: Give an alarm if measured rpm is under the min threshold (RW)
|
||||
Bit 3: Beep if alarm (RW)
|
||||
Bit 7: Shutdown if alarm persist for more then 4 seconds (RW)
|
||||
Bit 7: Shutdown if alarm persist for more than 4 seconds (RW)
|
||||
|
||||
Byte 1:
|
||||
min threshold (scale as bank 0x26)
|
||||
|
@ -540,7 +540,7 @@ A client would issue an operation by:
|
||||
MSG_MORE should be set in msghdr::msg_flags on all but the last part of
|
||||
the request. Multiple requests may be made simultaneously.
|
||||
|
||||
If a call is intended to go to a destination other then the default
|
||||
If a call is intended to go to a destination other than the default
|
||||
specified through connect(), then msghdr::msg_name should be set on the
|
||||
first request message of that call.
|
||||
|
||||
|
@ -733,7 +733,7 @@ Changes from 20040920 to 20041018
|
||||
I/O completion path a little more, especially taking care of
|
||||
fast-pathing the non-error case. Also removes tons of dead
|
||||
members and defines from lpfc_scsi.h - e.g. lpfc_target is down
|
||||
to nothing more then the lpfc_nodelist pointer.
|
||||
to nothing more than the lpfc_nodelist pointer.
|
||||
* Added binary sysfs file to issue mbox commands
|
||||
* Replaced #if __BIG_ENDIAN with #if __BIG_ENDIAN_BITFIELD for
|
||||
compatibility with the user space applications.
|
||||
|
@ -105,7 +105,7 @@ void pt_regs_to_gdb_regs(unsigned long *gdb_regs, struct pt_regs *regs)
|
||||
* Extracts ebp, esp and eip values understandable by gdb from the values
|
||||
* saved by switch_to.
|
||||
* thread.esp points to ebp. flags and ebp are pushed in switch_to hence esp
|
||||
* prior to entering switch_to is 8 greater then the value that is saved.
|
||||
* prior to entering switch_to is 8 greater than the value that is saved.
|
||||
* If switch_to changes, change following code appropriately.
|
||||
*/
|
||||
void sleeping_thread_to_gdb_regs(unsigned long *gdb_regs, struct task_struct *p)
|
||||
|
@ -434,7 +434,7 @@ int __kprobes trampoline_probe_handler(struct kprobe *p, struct pt_regs *regs)
|
||||
/*
|
||||
* It is possible to have multiple instances associated with a given
|
||||
* task either because an multiple functions in the call path
|
||||
* have a return probe installed on them, and/or more then one return
|
||||
* have a return probe installed on them, and/or more than one return
|
||||
* return probe was registered for a target function.
|
||||
*
|
||||
* We can handle this because:
|
||||
|
@ -303,7 +303,7 @@ config M68KFPU_EMU_EXTRAPREC
|
||||
correct rounding, the emulator can (often) do the same but this
|
||||
extra calculation can cost quite some time, so you can disable
|
||||
it here. The emulator will then "only" calculate with a 64 bit
|
||||
mantissa and round slightly incorrect, what is more then enough
|
||||
mantissa and round slightly incorrect, what is more than enough
|
||||
for normal usage.
|
||||
|
||||
config M68KFPU_EMU_ONLY
|
||||
|
@ -148,7 +148,7 @@ int read_eeprom(char *buffer, int eeprom_size, int size)
|
||||
send_byte(W_HEADER);
|
||||
recv_ack();
|
||||
|
||||
/* EEPROM with size of more then 2K need two byte addressing */
|
||||
/* EEPROM with size of more than 2K need two byte addressing */
|
||||
if (eeprom_size > 2048) {
|
||||
send_byte(0x00);
|
||||
recv_ack();
|
||||
|
@ -316,7 +316,7 @@ static int __kprobes trampoline_probe_handler(struct kprobe *p,
|
||||
/*
|
||||
* It is possible to have multiple instances associated with a given
|
||||
* task either because an multiple functions in the call path
|
||||
* have a return probe installed on them, and/or more then one return
|
||||
* have a return probe installed on them, and/or more than one return
|
||||
* return probe was registered for a target function.
|
||||
*
|
||||
* We can handle this because:
|
||||
|
@ -49,7 +49,7 @@ void set_spu_profiling_frequency(unsigned int freq_khz, unsigned int cycles_rese
|
||||
* of precision. This is close enough for the purpose at hand.
|
||||
*
|
||||
* The value of the timeout should be small enough that the hw
|
||||
* trace buffer will not get more then about 1/3 full for the
|
||||
* trace buffer will not get more than about 1/3 full for the
|
||||
* maximum user specified (the LFSR value) hw sampling frequency.
|
||||
* This is to ensure the trace buffer will never fill even if the
|
||||
* kernel thread scheduling varies under a heavy system load.
|
||||
|
@ -299,7 +299,7 @@ config WARN_STACK
|
||||
This option enables the compiler options -mwarn-framesize and
|
||||
-mwarn-dynamicstack. If the compiler supports these options it
|
||||
will generate warnings for function which either use alloca or
|
||||
create a stack frame bigger then CONFIG_WARN_STACK_SIZE.
|
||||
create a stack frame bigger than CONFIG_WARN_STACK_SIZE.
|
||||
|
||||
Say N if you are unsure.
|
||||
|
||||
|
@ -381,7 +381,7 @@ static int __kprobes trampoline_probe_handler(struct kprobe *p,
|
||||
/*
|
||||
* It is possible to have multiple instances associated with a given
|
||||
* task either because an multiple functions in the call path
|
||||
* have a return probe installed on them, and/or more then one return
|
||||
* have a return probe installed on them, and/or more than one return
|
||||
* return probe was registered for a target function.
|
||||
*
|
||||
* We can handle this because:
|
||||
|
@ -517,7 +517,7 @@ int __kprobes trampoline_probe_handler(struct kprobe *p, struct pt_regs *regs)
|
||||
/*
|
||||
* It is possible to have multiple instances associated with a given
|
||||
* task either because an multiple functions in the call path
|
||||
* have a return probe installed on them, and/or more then one return
|
||||
* have a return probe installed on them, and/or more than one return
|
||||
* return probe was registered for a target function.
|
||||
*
|
||||
* We can handle this because:
|
||||
|
@ -694,7 +694,7 @@ static __used __kprobes void *trampoline_handler(struct pt_regs *regs)
|
||||
/*
|
||||
* It is possible to have multiple instances associated with a given
|
||||
* task either because multiple functions in the call path have
|
||||
* return probes installed on them, and/or more then one
|
||||
* return probes installed on them, and/or more than one
|
||||
* return probe was registered for a target function.
|
||||
*
|
||||
* We can handle this because:
|
||||
|
@ -252,7 +252,7 @@ EXPORT_SYMBOL_GPL(geode_mfgpt_alloc_timer);
|
||||
/*
|
||||
* The MFPGT timers on the CS5536 provide us with suitable timers to use
|
||||
* as clock event sources - not as good as a HPET or APIC, but certainly
|
||||
* better then the PIT. This isn't a general purpose MFGPT driver, but
|
||||
* better than the PIT. This isn't a general purpose MFGPT driver, but
|
||||
* a simplified one designed specifically to act as a clock event source.
|
||||
* For full details about the MFGPT, please consult the CS5536 data sheet.
|
||||
*/
|
||||
|
@ -75,7 +75,7 @@ static const u8 FSCHMD_REG_VOLT[3] = { 0x45, 0x42, 0x48 };
|
||||
|
||||
/* minimum pwm at which the fan is driven (pwm can by increased depending on
|
||||
the temp. Notice that for the scy some fans share there minimum speed.
|
||||
Also notice that with the scy the sensor order is different then with the
|
||||
Also notice that with the scy the sensor order is different than with the
|
||||
other chips, this order was in the 2.4 driver and kept for consistency. */
|
||||
static const u8 FSCHMD_REG_FAN_MIN[5][6] = {
|
||||
{ 0x55, 0x65 }, /* pos */
|
||||
|
@ -367,7 +367,7 @@ int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata)
|
||||
if (err)
|
||||
goto out;
|
||||
} else {
|
||||
/* Can't be smaller then the number of outstanding CQEs */
|
||||
/* Can't be smaller than the number of outstanding CQEs */
|
||||
outst_cqe = mlx4_ib_get_outstanding_cqes(cq);
|
||||
if (entries < outst_cqe + 1) {
|
||||
err = 0;
|
||||
|
@ -390,7 +390,7 @@ static int i2o_scsi_reply(struct i2o_controller *c, u32 m,
|
||||
* @i2o_dev: the I2O device which was added
|
||||
*
|
||||
* If a I2O device is added we catch the notification, because I2O classes
|
||||
* other then SCSI peripheral will not be received through
|
||||
* other than SCSI peripheral will not be received through
|
||||
* i2o_scsi_probe().
|
||||
*/
|
||||
static void i2o_scsi_notify_device_add(struct i2o_device *i2o_dev)
|
||||
|
@ -34,7 +34,7 @@
|
||||
* aperture size, not the dram size, and the V370PDC supplies no
|
||||
* other method for memory size discovery. This problem is
|
||||
* mostly only relevant when compiled as a module, as the
|
||||
* unloading of the module with an aperture size smaller then
|
||||
* unloading of the module with an aperture size smaller than
|
||||
* the ram will cause the driver to detect the onboard memory
|
||||
* size to be equal to the aperture size when the module is
|
||||
* reloaded. Soooo, to help, the module supports an msize
|
||||
|
@ -717,7 +717,7 @@ write_error:
|
||||
* to the real data size, although the @buf buffer has to contain the
|
||||
* alignment. In all other cases, @len has to be aligned.
|
||||
*
|
||||
* It is prohibited to write more then once to logical eraseblocks of static
|
||||
* It is prohibited to write more than once to logical eraseblocks of static
|
||||
* volumes. This function returns zero in case of success and a negative error
|
||||
* code in case of failure.
|
||||
*/
|
||||
|
@ -465,7 +465,7 @@ out:
|
||||
* This function synchronously erases physical eraseblock @pnum. If @torture
|
||||
* flag is not zero, the physical eraseblock is checked by means of writing
|
||||
* different patterns to it and reading them back. If the torturing is enabled,
|
||||
* the physical eraseblock is erased more then once.
|
||||
* the physical eraseblock is erased more than once.
|
||||
*
|
||||
* This function returns the number of erasures made in case of success, %-EIO
|
||||
* if the erasure failed or the torturing test failed, and other negative error
|
||||
|
@ -478,7 +478,7 @@ int ubi_scan_add_used(struct ubi_device *ubi, struct ubi_scan_info *si,
|
||||
return 0;
|
||||
} else {
|
||||
/*
|
||||
* This logical eraseblock is older then the one found
|
||||
* This logical eraseblock is older than the one found
|
||||
* previously.
|
||||
*/
|
||||
if (cmp_res & 4)
|
||||
|
@ -135,7 +135,7 @@ enum {
|
||||
* The erase counter header takes 64 bytes and has a plenty of unused space for
|
||||
* future usage. The unused fields are zeroed. The @version field is used to
|
||||
* indicate the version of UBI implementation which is supposed to be able to
|
||||
* work with this UBI image. If @version is greater then the current UBI
|
||||
* work with this UBI image. If @version is greater than the current UBI
|
||||
* version, the image is rejected. This may be useful in future if something
|
||||
* is changed radically. This field is duplicated in the volume identifier
|
||||
* header.
|
||||
@ -187,7 +187,7 @@ struct ubi_ec_hdr {
|
||||
* (sequence number) is used to distinguish between older and newer versions of
|
||||
* logical eraseblocks.
|
||||
*
|
||||
* There are 2 situations when there may be more then one physical eraseblock
|
||||
* There are 2 situations when there may be more than one physical eraseblock
|
||||
* corresponding to the same logical eraseblock, i.e., having the same @vol_id
|
||||
* and @lnum values in the volume identifier header. Suppose we have a logical
|
||||
* eraseblock L and it is mapped to the physical eraseblock P.
|
||||
|
@ -577,7 +577,7 @@ static int init_volumes(struct ubi_device *ubi, const struct ubi_scan_info *si,
|
||||
if (vtbl[i].flags & UBI_VTBL_AUTORESIZE_FLG) {
|
||||
/* Auto re-size flag may be set only for one volume */
|
||||
if (ubi->autoresize_vol_id != -1) {
|
||||
ubi_err("more then one auto-resize volume (%d "
|
||||
ubi_err("more than one auto-resize volume (%d "
|
||||
"and %d)", ubi->autoresize_vol_id, i);
|
||||
kfree(vol);
|
||||
return -EINVAL;
|
||||
|
@ -128,7 +128,7 @@
|
||||
* situation when the picked physical eraseblock is constantly erased after the
|
||||
* data is written to it. So, we have a constant which limits the highest erase
|
||||
* counter of the free physical eraseblock to pick. Namely, the WL sub-system
|
||||
* does not pick eraseblocks with erase counter greater then the lowest erase
|
||||
* does not pick eraseblocks with erase counter greater than the lowest erase
|
||||
* counter plus %WL_FREE_MAX_DIFF.
|
||||
*/
|
||||
#define WL_FREE_MAX_DIFF (2*UBI_WL_THRESHOLD)
|
||||
@ -917,7 +917,7 @@ static int ensure_wear_leveling(struct ubi_device *ubi)
|
||||
/*
|
||||
* We schedule wear-leveling only if the difference between the
|
||||
* lowest erase counter of used physical eraseblocks and a high
|
||||
* erase counter of free physical eraseblocks is greater then
|
||||
* erase counter of free physical eraseblocks is greater than
|
||||
* %UBI_WL_THRESHOLD.
|
||||
*/
|
||||
e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
|
||||
|
@ -3359,7 +3359,7 @@ static u8 bnx2x_format_ver(u32 num, u8 *str, u16 len)
|
||||
u8 shift = 8*4;
|
||||
u8 digit;
|
||||
if (len < 10) {
|
||||
/* Need more then 10chars for this format */
|
||||
/* Need more than 10chars for this format */
|
||||
*str_ptr = '\0';
|
||||
return -EINVAL;
|
||||
}
|
||||
|
@ -6758,7 +6758,7 @@ static s32 e1000_get_cable_length(struct e1000_hw *hw, u16 *min_length,
|
||||
* returns: - E1000_ERR_XXX
|
||||
* E1000_SUCCESS
|
||||
*
|
||||
* For phy's older then IGP, this function simply reads the polarity bit in the
|
||||
* For phy's older than IGP, this function simply reads the polarity bit in the
|
||||
* Phy Status register. For IGP phy's, this bit is valid only if link speed is
|
||||
* 10 Mbps. If the link speed is 100 Mbps there is no polarity so this bit will
|
||||
* return 0. If the link speed is 1000 Mbps the polarity status is in the
|
||||
@ -6834,7 +6834,7 @@ static s32 e1000_check_polarity(struct e1000_hw *hw,
|
||||
* returns: - E1000_ERR_XXX
|
||||
* E1000_SUCCESS
|
||||
*
|
||||
* For phy's older then IGP, this function reads the Downshift bit in the Phy
|
||||
* For phy's older than IGP, this function reads the Downshift bit in the Phy
|
||||
* Specific Status register. For IGP phy's, it reads the Downgrade bit in the
|
||||
* Link Health register. In IGP this bit is latched high, so the driver must
|
||||
* read it immediately after link is established.
|
||||
|
@ -75,7 +75,7 @@ struct slip {
|
||||
unsigned long tx_errors; /* Planned stuff */
|
||||
unsigned long rx_dropped; /* No memory for skb */
|
||||
unsigned long tx_dropped; /* When MTU change */
|
||||
unsigned long rx_over_errors; /* Frame bigger then SLIP buf. */
|
||||
unsigned long rx_over_errors; /* Frame bigger than SLIP buf. */
|
||||
#ifdef SL_INCLUDE_CSLIP
|
||||
unsigned long tx_compressed;
|
||||
unsigned long rx_compressed;
|
||||
|
@ -1004,7 +1004,7 @@ static inline void bdx_rxdb_free_elem(struct rxdb *db, int n)
|
||||
* skb for rx. It assumes that Rx is desabled in HW
|
||||
* funcs are grouped for better cache usage
|
||||
*
|
||||
* RxD fifo is smaller then RxF fifo by design. Upon high load, RxD will be
|
||||
* RxD fifo is smaller than RxF fifo by design. Upon high load, RxD will be
|
||||
* filled and packets will be dropped by nic without getting into host or
|
||||
* cousing interrupt. Anyway, in that condition, host has no chance to proccess
|
||||
* all packets, but dropping in nic is cheaper, since it takes 0 cpu cycles
|
||||
@ -1826,7 +1826,7 @@ static void bdx_tx_free(struct bdx_priv *priv)
|
||||
*
|
||||
* Pushes desc to TxD fifo and overlaps it if needed.
|
||||
* NOTE: this func does not check for available space. this is responsibility
|
||||
* of the caller. Neither does it check that data size is smaller then
|
||||
* of the caller. Neither does it check that data size is smaller than
|
||||
* fifo size.
|
||||
*/
|
||||
static void bdx_tx_push_desc(struct bdx_priv *priv, void *data, int size)
|
||||
|
@ -3064,7 +3064,7 @@ static int smctr_load_node_addr(struct net_device *dev)
|
||||
* will consequently cause a timeout.
|
||||
*
|
||||
* NOTE 1: If the monitor_state is MS_BEACON_TEST_STATE, all transmit
|
||||
* queues other then the one used for the lobe_media_test should be
|
||||
* queues other than the one used for the lobe_media_test should be
|
||||
* disabled.!?
|
||||
*
|
||||
* NOTE 2: If the monitor_state is MS_BEACON_TEST_STATE and the receive_mask
|
||||
|
@ -1332,7 +1332,7 @@ static int ipw2100_power_cycle_adapter(struct ipw2100_priv *priv)
|
||||
IPW_AUX_HOST_RESET_REG_STOP_MASTER);
|
||||
|
||||
/* Step 2. Wait for stop Master Assert
|
||||
* (not more then 50us, otherwise ret error */
|
||||
* (not more than 50us, otherwise ret error */
|
||||
i = 5;
|
||||
do {
|
||||
udelay(IPW_WAIT_RESET_MASTER_ASSERT_COMPLETE_DELAY);
|
||||
|
@ -184,8 +184,8 @@ void rt2x00crypto_rx_insert_iv(struct sk_buff *skb, unsigned int align,
|
||||
* Make room for new data, note that we increase both
|
||||
* headsize and tailsize when required. The tailsize is
|
||||
* only needed when ICV data needs to be inserted and
|
||||
* the padding is smaller then the ICV data.
|
||||
* When alignment requirements is greater then the
|
||||
* the padding is smaller than the ICV data.
|
||||
* When alignment requirements is greater than the
|
||||
* ICV data we must trim the skb to the correct size
|
||||
* because we need to remove the extra bytes.
|
||||
*/
|
||||
|
@ -236,7 +236,7 @@ struct strip {
|
||||
unsigned long tx_errors; /* Planned stuff */
|
||||
unsigned long rx_dropped; /* No memory for skb */
|
||||
unsigned long tx_dropped; /* When MTU change */
|
||||
unsigned long rx_over_errors; /* Frame bigger then STRIP buf. */
|
||||
unsigned long rx_over_errors; /* Frame bigger than STRIP buf. */
|
||||
|
||||
unsigned long pps_timer; /* Timer to determine pps */
|
||||
unsigned long rx_pps_count; /* Counter to determine pps */
|
||||
|
@ -535,8 +535,8 @@ static int dasd_eer_open(struct inode *inp, struct file *filp)
|
||||
eerb->buffer_page_count > INT_MAX / PAGE_SIZE) {
|
||||
kfree(eerb);
|
||||
MESSAGE(KERN_WARNING, "can't open device since module "
|
||||
"parameter eer_pages is smaller then 1 or"
|
||||
" bigger then %d", (int)(INT_MAX / PAGE_SIZE));
|
||||
"parameter eer_pages is smaller than 1 or"
|
||||
" bigger than %d", (int)(INT_MAX / PAGE_SIZE));
|
||||
unlock_kernel();
|
||||
return -EINVAL;
|
||||
}
|
||||
|
@ -427,7 +427,7 @@ static int vmlogrdr_receive_data(struct vmlogrdr_priv_t *priv)
|
||||
buffer = priv->buffer + sizeof(int);
|
||||
}
|
||||
/*
|
||||
* If the record is bigger then our buffer, we receive only
|
||||
* If the record is bigger than our buffer, we receive only
|
||||
* a part of it. We can get the rest later.
|
||||
*/
|
||||
if (iucv_data_count > NET_BUFFER_SIZE)
|
||||
@ -437,7 +437,7 @@ static int vmlogrdr_receive_data(struct vmlogrdr_priv_t *priv)
|
||||
0, buffer, iucv_data_count,
|
||||
&priv->residual_length);
|
||||
spin_unlock_bh(&priv->priv_lock);
|
||||
/* An rc of 5 indicates that the record was bigger then
|
||||
/* An rc of 5 indicates that the record was bigger than
|
||||
* the buffer, which is OK for us. A 9 indicates that the
|
||||
* record was purged befor we could receive it.
|
||||
*/
|
||||
|
@ -1964,10 +1964,10 @@ lpfc_set_disctmo(struct lpfc_vport *vport)
|
||||
uint32_t tmo;
|
||||
|
||||
if (vport->port_state == LPFC_LOCAL_CFG_LINK) {
|
||||
/* For FAN, timeout should be greater then edtov */
|
||||
/* For FAN, timeout should be greater than edtov */
|
||||
tmo = (((phba->fc_edtov + 999) / 1000) + 1);
|
||||
} else {
|
||||
/* Normal discovery timeout should be > then ELS/CT timeout
|
||||
/* Normal discovery timeout should be > than ELS/CT timeout
|
||||
* FC spec states we need 3 * ratov for CT requests
|
||||
*/
|
||||
tmo = ((phba->fc_ratov * 3) + 3);
|
||||
|
@ -420,7 +420,7 @@ lpfc_sli_next_iocb_slot (struct lpfc_hba *phba, struct lpfc_sli_ring *pring)
|
||||
if (unlikely(pring->local_getidx >= max_cmd_idx)) {
|
||||
lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
|
||||
"0315 Ring %d issue: portCmdGet %d "
|
||||
"is bigger then cmd ring %d\n",
|
||||
"is bigger than cmd ring %d\n",
|
||||
pring->ringno,
|
||||
pring->local_getidx, max_cmd_idx);
|
||||
|
||||
@ -1628,12 +1628,12 @@ lpfc_sli_rsp_pointers_error(struct lpfc_hba *phba, struct lpfc_sli_ring *pring)
|
||||
{
|
||||
struct lpfc_pgp *pgp = &phba->port_gp[pring->ringno];
|
||||
/*
|
||||
* Ring <ringno> handler: portRspPut <portRspPut> is bigger then
|
||||
* Ring <ringno> handler: portRspPut <portRspPut> is bigger than
|
||||
* rsp ring <portRspMax>
|
||||
*/
|
||||
lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
|
||||
"0312 Ring %d handler: portRspPut %d "
|
||||
"is bigger then rsp ring %d\n",
|
||||
"is bigger than rsp ring %d\n",
|
||||
pring->ringno, le32_to_cpu(pgp->rspPutInx),
|
||||
pring->numRiocb);
|
||||
|
||||
@ -2083,12 +2083,12 @@ lpfc_sli_handle_slow_ring_event(struct lpfc_hba *phba,
|
||||
portRspPut = le32_to_cpu(pgp->rspPutInx);
|
||||
if (portRspPut >= portRspMax) {
|
||||
/*
|
||||
* Ring <ringno> handler: portRspPut <portRspPut> is bigger then
|
||||
* Ring <ringno> handler: portRspPut <portRspPut> is bigger than
|
||||
* rsp ring <portRspMax>
|
||||
*/
|
||||
lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
|
||||
"0303 Ring %d handler: portRspPut %d "
|
||||
"is bigger then rsp ring %d\n",
|
||||
"is bigger than rsp ring %d\n",
|
||||
pring->ringno, portRspPut, portRspMax);
|
||||
|
||||
phba->link_state = LPFC_HBA_ERROR;
|
||||
|
@ -1203,7 +1203,7 @@ static void e100_disable_txdma_channel(struct e100_serial *info)
|
||||
unsigned long flags;
|
||||
|
||||
/* Disable output DMA channel for the serial port in question
|
||||
* ( set to something other then serialX)
|
||||
* ( set to something other than serialX)
|
||||
*/
|
||||
local_irq_save(flags);
|
||||
DFLOW(DEBUG_LOG(info->line, "disable_txdma_channel %i\n", info->line));
|
||||
@ -1266,7 +1266,7 @@ static void e100_disable_rxdma_channel(struct e100_serial *info)
|
||||
unsigned long flags;
|
||||
|
||||
/* Disable input DMA channel for the serial port in question
|
||||
* ( set to something other then serialX)
|
||||
* ( set to something other than serialX)
|
||||
*/
|
||||
local_irq_save(flags);
|
||||
if (info->line == 0) {
|
||||
|
@ -1332,7 +1332,7 @@ static void vgacon_save_screen(struct vc_data *c)
|
||||
c->vc_y = screen_info.orig_y;
|
||||
}
|
||||
|
||||
/* We can't copy in more then the size of the video buffer,
|
||||
/* We can't copy in more than the size of the video buffer,
|
||||
* or we'll be copying in VGA BIOS */
|
||||
|
||||
if (!vga_is_gfx)
|
||||
|
@ -854,7 +854,7 @@ static int o2hb_thread(void *data)
|
||||
|
||||
while (!kthread_should_stop() && !reg->hr_unclean_stop) {
|
||||
/* We track the time spent inside
|
||||
* o2hb_do_disk_heartbeat so that we avoid more then
|
||||
* o2hb_do_disk_heartbeat so that we avoid more than
|
||||
* hr_timeout_ms between disk writes. On busy systems
|
||||
* this should result in a heartbeat which is less
|
||||
* likely to time itself out. */
|
||||
|
@ -9,7 +9,7 @@
|
||||
|
||||
/*
|
||||
* Logic: we've got two memory sums for each process, "shared", and
|
||||
* "non-shared". Shared memory may get counted more then once, for
|
||||
* "non-shared". Shared memory may get counted more than once, for
|
||||
* each process that owns it. Non-shared memory is counted
|
||||
* accurately.
|
||||
*/
|
||||
|
@ -40,7 +40,7 @@ config UBIFS_FS_ZLIB
|
||||
depends on UBIFS_FS
|
||||
default y
|
||||
help
|
||||
Zlib copresses better then LZO but it is slower. Say 'Y' if unsure.
|
||||
Zlib compresses better than LZO but it is slower. Say 'Y' if unsure.
|
||||
|
||||
# Debugging-related stuff
|
||||
config UBIFS_FS_DEBUG
|
||||
|
@ -142,7 +142,7 @@ static long long get_liability(struct ubifs_info *c)
|
||||
*
|
||||
* This function is called when an operation cannot be budgeted because there
|
||||
* is supposedly no free space. But in most cases there is some free space:
|
||||
* o budgeting is pessimistic, so it always budgets more then it is actually
|
||||
* o budgeting is pessimistic, so it always budgets more than it is actually
|
||||
* needed, so shrinking the liability is one way to make free space - the
|
||||
* cached data will take less space then it was budgeted for;
|
||||
* o GC may turn some dark space into free space (budgeting treats dark space
|
||||
@ -606,7 +606,7 @@ void ubifs_release_budget(struct ubifs_info *c, struct ubifs_budget_req *req)
|
||||
* @c: UBIFS file-system description object
|
||||
*
|
||||
* This function converts budget which was allocated for a new page of data to
|
||||
* the budget of changing an existing page of data. The latter is smaller then
|
||||
* the budget of changing an existing page of data. The latter is smaller than
|
||||
* the former, so this function only does simple re-calculation and does not
|
||||
* involve any write-back.
|
||||
*/
|
||||
|
@ -45,7 +45,7 @@
|
||||
#define SMALL_NODE_WM UBIFS_MAX_DENT_NODE_SZ
|
||||
|
||||
/*
|
||||
* GC may need to move more then one LEB to make progress. The below constants
|
||||
* GC may need to move more than one LEB to make progress. The below constants
|
||||
* define "soft" and "hard" limits on the number of LEBs the garbage collector
|
||||
* may move.
|
||||
*/
|
||||
|
@ -191,7 +191,7 @@ again:
|
||||
if (wbuf->lnum != -1 && avail >= len) {
|
||||
/*
|
||||
* Someone else has switched the journal head and we have
|
||||
* enough space now. This happens when more then one process is
|
||||
* enough space now. This happens when more than one process is
|
||||
* trying to write to the same journal head at the same time.
|
||||
*/
|
||||
dbg_jnl("return LEB %d back, already have LEB %d:%d",
|
||||
|
@ -151,7 +151,7 @@ static int shrink_tnc(struct ubifs_info *c, int nr, int age, int *contention)
|
||||
* @contention: if any contention, this is set to %1
|
||||
*
|
||||
* This function walks the list of mounted UBIFS file-systems and frees clean
|
||||
* znodes which are older then @age, until at least @nr znodes are freed.
|
||||
* znodes which are older than @age, until at least @nr znodes are freed.
|
||||
* Returns the number of freed znodes.
|
||||
*/
|
||||
static int shrink_tnc_trees(int nr, int age, int *contention)
|
||||
|
@ -1348,7 +1348,7 @@ xfs_finish_flags(
|
||||
{
|
||||
int ronly = (mp->m_flags & XFS_MOUNT_RDONLY);
|
||||
|
||||
/* Fail a mount where the logbuf is smaller then the log stripe */
|
||||
/* Fail a mount where the logbuf is smaller than the log stripe */
|
||||
if (xfs_sb_version_haslogv2(&mp->m_sb)) {
|
||||
if (mp->m_logbsize <= 0 &&
|
||||
mp->m_sb.sb_logsunit > XLOG_BIG_RECORD_BSIZE) {
|
||||
|
@ -83,7 +83,7 @@ typedef enum {
|
||||
* @datbuf: data buffer - if NULL only oob data are read/written
|
||||
* @oobbuf: oob data buffer
|
||||
*
|
||||
* Note, it is allowed to read more then one OOB area at one go, but not write.
|
||||
* Note, it is allowed to read more than one OOB area at one go, but not write.
|
||||
* The interface assumes that the OOB write requests program only one page's
|
||||
* OOB area.
|
||||
*/
|
||||
|
@ -327,9 +327,9 @@ extern struct spi_master *spi_busnum_to_master(u16 busnum);
|
||||
* @tx_dma: DMA address of tx_buf, if @spi_message.is_dma_mapped
|
||||
* @rx_dma: DMA address of rx_buf, if @spi_message.is_dma_mapped
|
||||
* @len: size of rx and tx buffers (in bytes)
|
||||
* @speed_hz: Select a speed other then the device default for this
|
||||
* @speed_hz: Select a speed other than the device default for this
|
||||
* transfer. If 0 the default (from @spi_device) is used.
|
||||
* @bits_per_word: select a bits_per_word other then the device default
|
||||
* @bits_per_word: select a bits_per_word other than the device default
|
||||
* for this transfer. If 0 the default (from @spi_device) is used.
|
||||
* @cs_change: affects chipselect after this transfer completes
|
||||
* @delay_usecs: microseconds to delay after this transfer before
|
||||
|
@ -253,7 +253,7 @@ struct ubi_mkvol_req {
|
||||
*
|
||||
* Re-sizing is possible for both dynamic and static volumes. But while dynamic
|
||||
* volumes may be re-sized arbitrarily, static volumes cannot be made to be
|
||||
* smaller then the number of bytes they bear. To arbitrarily shrink a static
|
||||
* smaller than the number of bytes they bear. To arbitrarily shrink a static
|
||||
* volume, it must be wiped out first (by means of volume update operation with
|
||||
* zero number of bytes).
|
||||
*/
|
||||
|
@ -475,7 +475,7 @@ pid_t task_session_nr_ns(struct task_struct *tsk, struct pid_namespace *ns)
|
||||
EXPORT_SYMBOL(task_session_nr_ns);
|
||||
|
||||
/*
|
||||
* Used by proc to find the first pid that is greater then or equal to nr.
|
||||
* Used by proc to find the first pid that is greater than or equal to nr.
|
||||
*
|
||||
* If there is a pid at nr this function is exactly the same as find_pid_ns.
|
||||
*/
|
||||
|
@ -45,7 +45,7 @@
|
||||
*
|
||||
* The value 8 is somewhat carefully chosen, as anything
|
||||
* larger can result in overflows. NSEC_PER_JIFFY grows as
|
||||
* HZ shrinks, so values greater then 8 overflow 32bits when
|
||||
* HZ shrinks, so values greater than 8 overflow 32bits when
|
||||
* HZ=100.
|
||||
*/
|
||||
#define JIFFIES_SHIFT 8
|
||||
|
@ -141,8 +141,8 @@ void sctp_auth_destroy_keys(struct list_head *keys)
|
||||
/* Compare two byte vectors as numbers. Return values
|
||||
* are:
|
||||
* 0 - vectors are equal
|
||||
* < 0 - vector 1 is smaller then vector2
|
||||
* > 0 - vector 1 is greater then vector2
|
||||
* < 0 - vector 1 is smaller than vector2
|
||||
* > 0 - vector 1 is greater than vector2
|
||||
*
|
||||
* Algorithm is:
|
||||
* This is performed by selecting the numerically smaller key vector...
|
||||
|
@ -4268,9 +4268,9 @@ nomem:
|
||||
|
||||
/*
|
||||
* Handle a protocol violation when the chunk length is invalid.
|
||||
* "Invalid" length is identified as smaller then the minimal length a
|
||||
* "Invalid" length is identified as smaller than the minimal length a
|
||||
* given chunk can be. For example, a SACK chunk has invalid length
|
||||
* if it's length is set to be smaller then the size of sctp_sack_chunk_t.
|
||||
* if its length is set to be smaller than the size of sctp_sack_chunk_t.
|
||||
*
|
||||
* We inform the other end by sending an ABORT with a Protocol Violation
|
||||
* error code.
|
||||
@ -4300,7 +4300,7 @@ static sctp_disposition_t sctp_sf_violation_chunklen(
|
||||
|
||||
/*
|
||||
* Handle a protocol violation when the parameter length is invalid.
|
||||
* "Invalid" length is identified as smaller then the minimal length a
|
||||
* "Invalid" length is identified as smaller than the minimal length a
|
||||
* given parameter can be.
|
||||
*/
|
||||
static sctp_disposition_t sctp_sf_violation_paramlen(
|
||||
|
@ -2717,7 +2717,7 @@ static int sctp_setsockopt_associnfo(struct sock *sk, char __user *optval, int o
|
||||
paths++;
|
||||
}
|
||||
|
||||
/* Only validate asocmaxrxt if we have more then
|
||||
/* Only validate asocmaxrxt if we have more than
|
||||
* one path/transport. We do this because path
|
||||
* retransmissions are only counted when we have more
|
||||
* then one path.
|
||||
|
@ -227,7 +227,7 @@ void sctp_tsnmap_skip(struct sctp_tsnmap *map, __u32 tsn)
|
||||
*/
|
||||
bitmap_zero(map->tsn_map, map->len);
|
||||
} else {
|
||||
/* If the gap is smaller then the map size,
|
||||
/* If the gap is smaller than the map size,
|
||||
* shift the map by 'gap' bits and update further.
|
||||
*/
|
||||
bitmap_shift_right(map->tsn_map, map->tsn_map, gap, map->len);
|
||||
|
@ -238,7 +238,7 @@ static void i_usX2Y_In04Int(struct urb *urb)
|
||||
send = 0;
|
||||
for (j = 0; j < URBS_AsyncSeq && !err; ++j)
|
||||
if (0 == usX2Y->AS04.urb[j]->status) {
|
||||
struct us428_p4out *p4out = us428ctls->p4out + send; // FIXME if more then 1 p4out is new, 1 gets lost.
|
||||
struct us428_p4out *p4out = us428ctls->p4out + send; // FIXME if more than 1 p4out is new, 1 gets lost.
|
||||
usb_fill_bulk_urb(usX2Y->AS04.urb[j], usX2Y->chip.dev,
|
||||
usb_sndbulkpipe(usX2Y->chip.dev, 0x04), &p4out->val.vol,
|
||||
p4out->type == eLT_Light ? sizeof(struct us428_lights) : 5,
|
||||
|
Loading…
Reference in New Issue
Block a user