cpu/intel: Wrap lines at 80 columns

Fix the following warning detected by checkpatch.pl:

WARNING: line over 80 characters

TEST=Build and run on Galileo Gen2

Change-Id: I74f25da5c53bd518189ce86817d6e3385b29c3b4
Signed-off-by: Lee Leahy <Leroy.P.Leahy@intel.com>
Reviewed-on: https://review.coreboot.org/18850
Tested-by: build bot (Jenkins)
Reviewed-by: Martin Roth <martinroth@google.com>
diff --git a/src/cpu/intel/common/common_init.c b/src/cpu/intel/common/common_init.c
index da7b826..44a316b 100644
--- a/src/cpu/intel/common/common_init.c
+++ b/src/cpu/intel/common/common_init.c
@@ -40,7 +40,8 @@
 	msr = rdmsr(IA32_FEATURE_CONTROL);
 
 	if (msr.lo & (1 << 0)) {
-		printk(BIOS_ERR, "VMX is locked, so %s will do nothing\n", __func__);
+		printk(BIOS_ERR, "VMX is locked, so %s will do nothing\n",
+			__func__);
 		/* VMX locked. If we set it again we get an illegal
 		 * instruction
 		 */
@@ -66,6 +67,7 @@
 		wrmsr(IA32_FEATURE_CONTROL, msr);
 	}
 
-	printk(BIOS_DEBUG, "VMX status: %s, %s\n", enable ? "enabled" : "disabled",
+	printk(BIOS_DEBUG, "VMX status: %s, %s\n",
+		enable ? "enabled" : "disabled",
 		lock ? "locked" : "unlocked");
 }
diff --git a/src/cpu/intel/fsp_model_406dx/bootblock.c b/src/cpu/intel/fsp_model_406dx/bootblock.c
index ee4cfac..a208ec9 100644
--- a/src/cpu/intel/fsp_model_406dx/bootblock.c
+++ b/src/cpu/intel/fsp_model_406dx/bootblock.c
@@ -33,11 +33,13 @@
 {
 
 	/*
-	 * Check if INIT# is asserted by port 0xCF9 and whether RCBA has been set.
-	 * If either is true, then this is a warm reset so execute a Hard Reset
+	 * Check if INIT# is asserted by port 0xCF9 and whether RCBA has been
+	 * set.   If either is true, then this is a warm reset so execute a
+	 * Hard Reset
 	 */
 	if ((inb(0xcf9) == 0x04) ||
-			(pci_io_read_config32(SOC_LPC_DEV, RCBA) & RCBA_ENABLE)) {
+			(pci_io_read_config32(SOC_LPC_DEV, RCBA)
+			& RCBA_ENABLE)) {
 		outb(0x00, 0xcf9);
 		outb(0x06, 0xcf9);
 	}
diff --git a/src/cpu/intel/haswell/haswell_init.c b/src/cpu/intel/haswell/haswell_init.c
index 0f6be39..e850fd1 100644
--- a/src/cpu/intel/haswell/haswell_init.c
+++ b/src/cpu/intel/haswell/haswell_init.c
@@ -419,7 +419,8 @@
 	u8 power_limit_1_val;
 
 	if (power_limit_1_time >= ARRAY_SIZE(power_limit_time_sec_to_msr))
-		power_limit_1_time = ARRAY_SIZE(power_limit_time_sec_to_msr) - 1;
+		power_limit_1_time = ARRAY_SIZE(power_limit_time_sec_to_msr)
+		- 1;
 
 	if (!(msr.lo & PLATFORM_INFO_SET_TDP))
 		return;
diff --git a/src/cpu/intel/haswell/smmrelocate.c b/src/cpu/intel/haswell/smmrelocate.c
index e7a9ee1..5c88dfa 100644
--- a/src/cpu/intel/haswell/smmrelocate.c
+++ b/src/cpu/intel/haswell/smmrelocate.c
@@ -255,7 +255,8 @@
 	/* SMRR has 32-bits of valid address aligned to 4KiB. */
 	params->smrr_base.lo = (params->smram_base & rmask) | MTRR_TYPE_WRBACK;
 	params->smrr_base.hi = 0;
-	params->smrr_mask.lo = (~(tseg_size - 1) & rmask) | MTRR_PHYS_MASK_VALID;
+	params->smrr_mask.lo = (~(tseg_size - 1) & rmask)
+		| MTRR_PHYS_MASK_VALID;
 	params->smrr_mask.hi = 0;
 
 	/* The EMRR and UNCORE_EMRR are at IEDBASE + 2MiB */
@@ -266,7 +267,8 @@
 	 * on the number of physical address bits supported. */
 	params->emrr_base.lo = emrr_base | MTRR_TYPE_WRBACK;
 	params->emrr_base.hi = 0;
-	params->emrr_mask.lo = (~(emrr_size - 1) & rmask) | MTRR_PHYS_MASK_VALID;
+	params->emrr_mask.lo = (~(emrr_size - 1) & rmask)
+		| MTRR_PHYS_MASK_VALID;
 	params->emrr_mask.hi = (1 << (phys_bits - 32)) - 1;
 
 	/* UNCORE_EMRR has 39 bits of valid address aligned to 4KiB. */
diff --git a/src/cpu/intel/microcode/microcode.c b/src/cpu/intel/microcode/microcode.c
index 1ab3027..59c3b8a 100644
--- a/src/cpu/intel/microcode/microcode.c
+++ b/src/cpu/intel/microcode/microcode.c
@@ -209,7 +209,8 @@
 }
 
 #if ENV_RAMSTAGE
-__attribute__((weak)) int soc_skip_ucode_update(u32 currrent_patch_id, u32 new_patch_id)
+__attribute__((weak)) int soc_skip_ucode_update(u32 currrent_patch_id,
+	u32 new_patch_id)
 {
 	return 0;
 }
diff --git a/src/cpu/intel/model_1067x/model_1067x_init.c b/src/cpu/intel/model_1067x/model_1067x_init.c
index 9b7536d..01cb45d 100644
--- a/src/cpu/intel/model_1067x/model_1067x_init.c
+++ b/src/cpu/intel/model_1067x/model_1067x_init.c
@@ -83,7 +83,8 @@
 
 	/* Set Processor MWAIT IO BASE */
 	msr.hi = 0;
-	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff) << 16);
+	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff)
+		<< 16);
 	wrmsr(MSR_PMG_IO_BASE_ADDR, msr);
 
 	/* Set IO Capture Address */
diff --git a/src/cpu/intel/model_106cx/model_106cx_init.c b/src/cpu/intel/model_106cx/model_106cx_init.c
index d9ce736..b07fd0b 100644
--- a/src/cpu/intel/model_106cx/model_106cx_init.c
+++ b/src/cpu/intel/model_106cx/model_106cx_init.c
@@ -34,7 +34,8 @@
 
 	msr = rdmsr(MSR_PMG_CST_CONFIG_CONTROL);
 	msr.lo |= (1 << 15); // Lock configuration
-	msr.lo |= (1 << 10); // redirect IO-based CState transition requests to MWAIT
+	msr.lo |= (1 << 10); // redirect IO-based CState transition requests to
+			     // MWAIT
 	msr.lo &= ~(1 << 9); // Issue a single stop grant cycle upon stpclk
 	msr.lo &= ~7; msr.lo |= HIGHEST_CLEVEL; // support at most C3
 	// TODO Do we want Deep C4 and  Dynamic L2 shrinking?
@@ -43,13 +44,15 @@
 	/* Set Processor MWAIT IO BASE (P_BLK) */
 	msr.hi = 0;
 	// TODO Do we want PM1_BASE? Needs SMM?
-	//msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff) << 16);
+	//msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff)
+	//	 << 16);
 	msr.lo = ((PMB0_BASE + 4) & 0xffff);
 	wrmsr(MSR_PMG_IO_BASE_ADDR, msr);
 
 	/* set C_LVL controls */
 	msr.hi = 0;
-	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16); // -2 because LVL0+1 aren't counted
+	// -2 because LVL0+1 aren't counted
+	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16);
 	wrmsr(MSR_PMG_IO_CAPTURE_ADDR, msr);
 }
 
diff --git a/src/cpu/intel/model_6dx/model_6dx_init.c b/src/cpu/intel/model_6dx/model_6dx_init.c
index a18b00c..0a51914 100644
--- a/src/cpu/intel/model_6dx/model_6dx_init.c
+++ b/src/cpu/intel/model_6dx/model_6dx_init.c
@@ -40,8 +40,9 @@
 };
 
 static struct cpu_device_id cpu_table[] = {
-	{ X86_VENDOR_INTEL, 0x06D0 }, /* Pentium M on 90nm with 2MiB of L2 cache */
-	{ X86_VENDOR_INTEL, 0x06D6 }, /* Pentium M on 90nm with 2MiB of L2 cache */
+	/* Pentium M on 90nm with 2MiB of L2 cache */
+	{ X86_VENDOR_INTEL, 0x06D0 },
+	{ X86_VENDOR_INTEL, 0x06D6 },
 	{ 0, 0 },
 };
 
diff --git a/src/cpu/intel/model_6ex/model_6ex_init.c b/src/cpu/intel/model_6ex/model_6ex_init.c
index 8381c70..ff9284f 100644
--- a/src/cpu/intel/model_6ex/model_6ex_init.c
+++ b/src/cpu/intel/model_6ex/model_6ex_init.c
@@ -47,12 +47,14 @@
 
 	/* Set Processor MWAIT IO BASE (P_BLK) */
 	msr.hi = 0;
-	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff) << 16);
+	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff)
+		<< 16);
 	wrmsr(MSR_PMG_IO_BASE_ADDR, msr);
 
 	/* Set C_LVL controls and IO Capture Address */
 	msr.hi = 0;
-	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16); // -2 because LVL0+1 aren't counted
+	// -2 because LVL0+1 aren't counted
+	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16);
 	wrmsr(MSR_PMG_IO_CAPTURE_ADDR, msr);
 }
 
diff --git a/src/cpu/intel/model_6fx/model_6fx_init.c b/src/cpu/intel/model_6fx/model_6fx_init.c
index 4c13f12..2a04a4e 100644
--- a/src/cpu/intel/model_6fx/model_6fx_init.c
+++ b/src/cpu/intel/model_6fx/model_6fx_init.c
@@ -48,12 +48,14 @@
 
 	/* Set Processor MWAIT IO BASE (P_BLK) */
 	msr.hi = 0;
-	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff) << 16);
+	msr.lo = ((PMB0_BASE + 4) & 0xffff) | (((PMB1_BASE + 9) & 0xffff)
+		<< 16);
 	wrmsr(MSR_PMG_IO_BASE_ADDR, msr);
 
 	/* Set C_LVL controls and IO Capture Address */
 	msr.hi = 0;
-	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16); // -2 because LVL0+1 aren't counted
+	// -2 because LVL0+1 aren't counted
+	msr.lo = (PMB0_BASE + 4) | ((HIGHEST_CLEVEL - 2) << 16);
 	wrmsr(MSR_PMG_IO_CAPTURE_ADDR, msr);
 }
 
diff --git a/src/cpu/intel/slot_1/l2_cache.c b/src/cpu/intel/slot_1/l2_cache.c
index b7b9719..05eaeb3 100644
--- a/src/cpu/intel/slot_1/l2_cache.c
+++ b/src/cpu/intel/slot_1/l2_cache.c
@@ -169,7 +169,8 @@
 		msr = rdmsr(EBL_CR_POWERON);
 		/* Get clock multiplier and FSB frequency.
 		 * Multiplier is in [25:22].
-		 * FSB is in [19:18] in Katmai, [19] in Deschutes ([18] is zero for them).
+		 * FSB is in [19:18] in Katmai, [19] in Deschutes ([18] is zero
+		 * for them).
 		 */
 		eax = msr.lo >> 18;
 		if (signature == 0x650) {
@@ -185,7 +186,9 @@
 		for (le = latency_table; le->key != eax; le++) {
 			/* Fail if we get to the end of the table */
 			if (le->key == 0xff) {
-				printk(BIOS_DEBUG, "Could not find key %02x in latency table\n", eax);
+				printk(BIOS_DEBUG,
+				   "Could not find key %02x in latency table\n",
+				   eax);
 				return -1;
 			}
 		}
@@ -281,7 +284,8 @@
 	} else
 		v2 &= 0x7;
 
-	/* This write has to be replicated to a number of places. Not sure what. */
+	/* This write has to be replicated to a number of places. Not sure what.
+	 */
 
 	for (i = 0; i < v2; i++) {
 
@@ -339,7 +343,8 @@
 /* Calculates the L2 cache size.
  *
  * Reference: Intel(R) 64 and IA-32 Architectures Software Developer's Manual
- *            Volume 3B: System Programming Guide, Part 2, Intel pub. 253669, pg. B-172.
+ *            Volume 3B: System Programming Guide, Part 2, Intel pub. 253669,
+ *            pg. B-172.
  *
  */
 int calculate_l2_cache_size(void)
@@ -357,8 +362,9 @@
 		bblcr3 = msr.lo & ~BBLCR3_L2_SIZE;
 		/*
 		 * Successively write in all the possible cache size per bank
-		 * into BBL_CR_CTL3[17:13], starting from 256KB (00001) to 4MB (10000),
-		 * and read the last value written and accepted by the cache.
+		 * into BBL_CR_CTL3[17:13], starting from 256KB (00001) to 4MB
+		 * (10000), and read the last value written and accepted by the
+		 * cache.
 		 *
 		 * No idea why these bits are writable at all.
 		 */
@@ -386,8 +392,9 @@
 
 		printk(BIOS_DEBUG, "Maximum cache mask is %x\n", cache_setting);
 
-		/* For now, BBL_CR_CTL3 has the highest cache "size" that register
-		 * will accept. Now we'll ping the cache and see where it wraps.
+		/* For now, BBL_CR_CTL3 has the highest cache "size" that
+		 * register will accept. Now we'll ping the cache and see where
+		 * it wraps.
 		 */
 
 		/* Write aaaaaaaa:aaaaaaaa to address 0 in the l2 cache.
@@ -493,7 +500,8 @@
 	return 0;
 }
 
-// L2 physical address range can be found from L2 control register 3, bits [2:0].
+// L2 physical address range can be found from L2 control register 3,
+// bits [2:0].
 int calculate_l2_physical_address_range(void)
 {
 	int r0, r3;
@@ -512,7 +520,8 @@
 	else
 		r3 &= 0x7;
 
-	printk(BIOS_DEBUG, "L2 Physical Address Range is %dM\n", (1 << r3) * 512);
+	printk(BIOS_DEBUG, "L2 Physical Address Range is %dM\n",
+		(1 << r3) * 512);
 
 	/* Shift into [22:20] to be saved into BBL_CR_CTL3. */
 	r3 = r3 << 20;
@@ -659,8 +668,9 @@
 		bblctl3.lo = eax;
 		wrmsr(BBL_CR_CTL3, bblctl3);
 
-		/* Write BBL_CR_CTL3[27:26] (reserved??) to bits [1:0] of L2 register 4.
-		 * Apparently all other bits must be preserved, hence these code.
+		/* Write BBL_CR_CTL3[27:26] (reserved??) to bits [1:0] of L2
+		 * register 4.  Apparently all other bits must be preserved,
+		 * hence these code.
 		 */
 
 		v = (calc_eax >> 26) & 0x3;
@@ -685,7 +695,8 @@
 	/* Read L2 register 0 */
 	v = read_l2(0);
 
-	/* If L2(0)[5] set (and can be read properly), enable CRTN and address parity
+	/* If L2(0)[5] set (and can be read properly), enable CRTN and address
+	 * parity
 	 */
 	if (v >= 0 && (v & 0x20)) {
 		bblctl3 = rdmsr(BBL_CR_CTL3);
@@ -700,7 +711,8 @@
 	set_l2_ecc();
 
 	if (calculate_l2_physical_address_range() != 0) {
-		printk(BIOS_ERR, "Failed to calculate L2 physical address range");
+		printk(BIOS_ERR,
+			"Failed to calculate L2 physical address range");
 		goto bad;
 	}
 
@@ -738,9 +750,10 @@
 
 		/* Update each way */
 
-		/* We're supposed to get L2 associativity from BBL_CR_CTL3[10:9].
-		 * But this code only applies to certain members of the P6 processor family
-		 * and since all P6 processors have 4-way L2 cache, we can safely assume
+		/* We're supposed to get L2 associativity from
+		 * BBL_CR_CTL3[10:9].  But this code only applies to certain
+		 * members of the P6 processor family and since all P6
+		 * processors have 4-way L2 cache, we can safely assume
 		 * 4 way for all cache operations.
 		 */
 
@@ -748,8 +761,10 @@
 			/* Send Tag Write w/Data Write (TWW) to L2 controller
 			 * MESI = Invalid
 			 */
-			if (signal_l2(cache_size, 0, 0, v, L2CMD_TWW | L2CMD_MESI_I) != 0) {
-				printk(BIOS_ERR, "Failed on signal_l2(%x, %x)\n",
+			if (signal_l2(cache_size, 0, 0, v, L2CMD_TWW
+				| L2CMD_MESI_I) != 0) {
+				printk(BIOS_ERR,
+					"Failed on signal_l2(%x, %x)\n",
 				       cache_size, v);
 				goto bad;
 			}
diff --git a/src/cpu/intel/smm/gen1/smmrelocate.c b/src/cpu/intel/smm/gen1/smmrelocate.c
index c3596d2..1da7a4b 100644
--- a/src/cpu/intel/smm/gen1/smmrelocate.c
+++ b/src/cpu/intel/smm/gen1/smmrelocate.c
@@ -133,7 +133,8 @@
 	/* SMRR has 32-bits of valid address aligned to 4KiB. */
 	params->smrr_base.lo = (params->smram_base & rmask) | MTRR_TYPE_WRBACK;
 	params->smrr_base.hi = 0;
-	params->smrr_mask.lo = (~(tseg_size - 1) & rmask) | MTRR_PHYS_MASK_VALID;
+	params->smrr_mask.lo = (~(tseg_size - 1) & rmask)
+		| MTRR_PHYS_MASK_VALID;
 	params->smrr_mask.hi = 0;
 }
 
@@ -226,12 +227,14 @@
 		       num_cpus, CONFIG_MAX_CPUS);
 	}
 
-	if (install_relocation_handler(apic_id_map, num_cpus, &smm_reloc_params)) {
+	if (install_relocation_handler(apic_id_map, num_cpus,
+		&smm_reloc_params)) {
 		printk(BIOS_CRIT, "SMM Relocation handler install failed.\n");
 		return -1;
 	}
 
-	if (install_permanent_handler(apic_id_map, num_cpus, &smm_reloc_params)) {
+	if (install_permanent_handler(apic_id_map, num_cpus,
+		&smm_reloc_params)) {
 		printk(BIOS_CRIT, "SMM Permanent handler install failed.\n");
 		return -1;
 	}