1 /* Smedia Glamo 336x/337x driver
3 * (C) 2007 by Openmoko, Inc.
4 * Author: Harald Welte <laforge@openmoko.org>
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License as
9 * published by the Free Software Foundation; either version 2 of
10 * the License, or (at your option) any later version.
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 59 Temple Place, Suite 330, Boston,
23 #include <linux/module.h>
24 #include <linux/kernel.h>
25 #include <linux/errno.h>
26 #include <linux/string.h>
28 #include <linux/delay.h>
30 #include <linux/init.h>
31 #include <linux/irq.h>
32 #include <linux/interrupt.h>
33 #include <linux/workqueue.h>
34 #include <linux/platform_device.h>
35 #include <linux/kernel_stat.h>
36 #include <linux/spinlock.h>
37 #include <linux/mfd/core.h>
38 #include <linux/mfd/glamo.h>
39 #include <linux/spi/glamo.h>
40 #include <linux/glamo-gpio.h>
41 #include <linux/glamofb.h>
44 #include <asm/div64.h>
48 #include "glamo-regs.h"
49 #include "glamo-core.h"
51 #define GLAMO_MEM_REFRESH_COUNT 0x100
54 * Glamo internal settings
56 * We run the memory interface from the faster PLLB on 2.6.28 kernels and
57 * above. Couple of GTA02 users report trouble with memory bus when they
58 * upgraded from 2.6.24. So this parameter allows reversion to 2.6.24
59 * scheme if their Glamo chip needs it.
61 * you can override the faster default on kernel commandline using
63 * glamo3362.slow_memory=1
68 static int slow_memory
= 0;
69 module_param(slow_memory
, int, 0644);
78 struct reg_range reg_range
[] = {
79 { 0x0000, 0x76, "General", 1 },
80 { 0x0200, 0x18, "Host Bus", 1 },
81 { 0x0300, 0x38, "Memory", 1 },
82 /* { 0x0400, 0x100, "Sensor", 0 }, */
83 /* { 0x0500, 0x300, "ISP", 0 }, */
84 /* { 0x0800, 0x400, "JPEG", 0 }, */
85 /* { 0x0c00, 0xcc, "MPEG", 0 }, */
86 { 0x1100, 0xb2, "LCD 1", 1 },
87 { 0x1200, 0x64, "LCD 2", 1 },
88 { 0x1400, 0x42, "MMC", 1 },
89 /* { 0x1500, 0x080, "MPU 0", 0 },
90 { 0x1580, 0x080, "MPU 1", 0 },
91 { 0x1600, 0x080, "Cmd Queue", 0 },
92 { 0x1680, 0x080, "RISC CPU", 0 },
93 { 0x1700, 0x400, "2D Unit", 0 },
94 { 0x1b00, 0x900, "3D Unit", 0 }, */
97 static inline void __reg_write(struct glamo_core
*glamo
,
98 u_int16_t reg
, u_int16_t val
)
100 writew(val
, glamo
->base
+ reg
);
103 static inline u_int16_t
__reg_read(struct glamo_core
*glamo
,
106 return readw(glamo
->base
+ reg
);
109 static void __reg_set_bit_mask(struct glamo_core
*glamo
,
110 u_int16_t reg
, u_int16_t mask
,
117 tmp
= __reg_read(glamo
, reg
);
120 __reg_write(glamo
, reg
, tmp
);
123 static void reg_set_bit_mask(struct glamo_core
*glamo
,
124 u_int16_t reg
, u_int16_t mask
,
127 spin_lock(&glamo
->lock
);
128 __reg_set_bit_mask(glamo
, reg
, mask
, val
);
129 spin_unlock(&glamo
->lock
);
132 static inline void __reg_set_bit(struct glamo_core
*glamo
,
133 u_int16_t reg
, u_int16_t bit
)
135 __reg_set_bit_mask(glamo
, reg
, bit
, 0xffff);
138 static inline void __reg_clear_bit(struct glamo_core
*glamo
,
139 u_int16_t reg
, u_int16_t bit
)
141 __reg_set_bit_mask(glamo
, reg
, bit
, 0);
144 /***********************************************************************
145 * resources of sibling devices
146 ***********************************************************************/
148 static struct resource glamo_fb_resources
[] = {
150 .name
= "glamo-fb-regs",
151 .start
= GLAMO_REGOFS_LCD
,
152 .end
= GLAMO_REGOFS_MMC
- 1,
153 .flags
= IORESOURCE_MEM
,
155 .name
= "glamo-fb-mem",
156 .start
= GLAMO_OFFSET_FB
,
157 .end
= GLAMO_OFFSET_FB
+ GLAMO_FB_SIZE
- 1,
158 .flags
= IORESOURCE_MEM
,
162 static struct resource glamo_mmc_resources
[] = {
164 .start
= GLAMO_REGOFS_MMC
,
165 .end
= GLAMO_REGOFS_MPROC0
- 1,
166 .flags
= IORESOURCE_MEM
168 .start
= IRQ_GLAMO_MMC
,
169 .end
= IRQ_GLAMO_MMC
,
170 .flags
= IORESOURCE_IRQ
,
171 }, { /* our data buffer for MMC transfers */
172 .start
= GLAMO_OFFSET_FB
+ GLAMO_FB_SIZE
,
173 .end
= GLAMO_OFFSET_FB
+ GLAMO_FB_SIZE
+
174 GLAMO_MMC_BUFFER_SIZE
- 1,
175 .flags
= IORESOURCE_MEM
185 static struct mfd_cell glamo_cells
[] = {
188 .num_resources
= ARRAY_SIZE(glamo_fb_resources
),
189 .resources
= glamo_fb_resources
,
193 .num_resources
= ARRAY_SIZE(glamo_mmc_resources
),
194 .resources
= glamo_mmc_resources
,
196 [GLAMO_CELL_SPI_GPIO
] = {
197 .name
= "glamo-spi-gpio",
202 /***********************************************************************
204 ***********************************************************************/
205 #define irq2glamo(x) (x - IRQ_GLAMO(0))
207 static void glamo_ack_irq(unsigned int irq
)
209 struct glamo_core
*glamo
= (struct glamo_core
*)get_irq_chip_data(irq
);
210 /* clear interrupt source */
211 __reg_write(glamo
, GLAMO_REG_IRQ_CLEAR
,
212 1 << irq2glamo(irq
));
215 static void glamo_mask_irq(unsigned int irq
)
217 struct glamo_core
*glamo
= (struct glamo_core
*)get_irq_chip_data(irq
);
220 /* clear bit in enable register */
221 tmp
= __reg_read(glamo
, GLAMO_REG_IRQ_ENABLE
);
222 tmp
&= ~(1 << irq2glamo(irq
));
223 __reg_write(glamo
, GLAMO_REG_IRQ_ENABLE
, tmp
);
226 static void glamo_unmask_irq(unsigned int irq
)
228 struct glamo_core
*glamo
= (struct glamo_core
*)get_irq_chip_data(irq
);
231 /* set bit in enable register */
232 tmp
= __reg_read(glamo
, GLAMO_REG_IRQ_ENABLE
);
233 tmp
|= (1 << irq2glamo(irq
));
234 __reg_write(glamo
, GLAMO_REG_IRQ_ENABLE
, tmp
);
237 static struct irq_chip glamo_irq_chip
= {
239 .ack
= glamo_ack_irq
,
240 .mask
= glamo_mask_irq
,
241 .unmask
= glamo_unmask_irq
,
244 static void glamo_irq_demux_handler(unsigned int irq
, struct irq_desc
*desc
)
246 struct glamo_core
*glamo
= get_irq_desc_chip_data(desc
);
247 desc
->status
&= ~(IRQ_REPLAY
| IRQ_WAITING
);
249 if (unlikely(desc
->status
& IRQ_INPROGRESS
)) {
250 desc
->status
|= (IRQ_PENDING
| IRQ_MASKED
);
251 desc
->chip
->mask(irq
);
252 desc
->chip
->ack(irq
);
255 kstat_incr_irqs_this_cpu(irq
, desc
);
257 desc
->chip
->ack(irq
);
258 desc
->status
|= IRQ_INPROGRESS
;
264 if (unlikely((desc
->status
&
265 (IRQ_PENDING
| IRQ_MASKED
| IRQ_DISABLED
)) ==
266 (IRQ_PENDING
| IRQ_MASKED
))) {
267 /* dealing with pending IRQ, unmasking */
268 desc
->chip
->unmask(irq
);
269 desc
->status
&= ~IRQ_MASKED
;
272 desc
->status
&= ~IRQ_PENDING
;
274 /* read IRQ status register */
275 irqstatus
= __reg_read(glamo
, GLAMO_REG_IRQ_STATUS
);
276 for (i
= 0; i
< 9; i
++)
277 if (irqstatus
& (1 << i
))
278 desc_handle_irq(IRQ_GLAMO(i
),
279 irq_desc
+IRQ_GLAMO(i
));
281 } while ((desc
->status
& (IRQ_PENDING
| IRQ_DISABLED
)) == IRQ_PENDING
);
283 desc
->status
&= ~IRQ_INPROGRESS
;
287 static ssize_t
regs_write(struct device
*dev
, struct device_attribute
*attr
,
288 const char *buf
, size_t count
)
290 unsigned long reg
= simple_strtoul(buf
, NULL
, 10);
291 struct glamo_core
*glamo
= dev_get_drvdata(dev
);
293 while (*buf
&& (*buf
!= ' '))
297 while (*buf
&& (*buf
== ' '))
302 printk(KERN_INFO
"reg 0x%02lX <-- 0x%04lX\n",
303 reg
, simple_strtoul(buf
, NULL
, 10));
305 __reg_write(glamo
, reg
, simple_strtoul(buf
, NULL
, 10));
310 static ssize_t
regs_read(struct device
*dev
, struct device_attribute
*attr
,
313 struct glamo_core
*glamo
= dev_get_drvdata(dev
);
317 spin_lock(&glamo
->lock
);
319 for (r
= 0; r
< ARRAY_SIZE(reg_range
); r
++) {
320 if (!reg_range
[r
].dump
)
323 end
+= sprintf(end
, "\n%s\n", reg_range
[r
].name
);
324 for (n
= reg_range
[r
].start
;
325 n
< reg_range
[r
].start
+ reg_range
[r
].count
; n
+= 2) {
326 if (((n1
++) & 7) == 0)
327 end
+= sprintf(end
, "\n%04X: ", n
);
328 end
+= sprintf(end
, "%04x ", __reg_read(glamo
, n
));
330 end
+= sprintf(end
, "\n");
336 spin_unlock(&glamo
->lock
);
341 static DEVICE_ATTR(regs
, 0644, regs_read
, regs_write
);
342 static struct attribute
*glamo_sysfs_entries
[] = {
346 static struct attribute_group glamo_attr_group
= {
348 .attrs
= glamo_sysfs_entries
,
353 /***********************************************************************
355 ***********************************************************************/
357 int __glamo_engine_enable(struct glamo_core
*glamo
, enum glamo_engine engine
)
360 case GLAMO_ENGINE_LCD
:
361 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
362 GLAMO_HOSTBUS2_MMIO_EN_LCD
,
363 GLAMO_HOSTBUS2_MMIO_EN_LCD
);
364 __reg_write(glamo
, GLAMO_REG_CLOCK_LCD
,
365 GLAMO_CLOCK_LCD_EN_M5CLK
|
366 GLAMO_CLOCK_LCD_EN_DHCLK
|
367 GLAMO_CLOCK_LCD_EN_DMCLK
|
368 GLAMO_CLOCK_LCD_EN_DCLK
|
369 GLAMO_CLOCK_LCD_DG_M5CLK
|
370 GLAMO_CLOCK_LCD_DG_DMCLK
);
371 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
372 GLAMO_CLOCK_GEN51_EN_DIV_DHCLK
|
373 GLAMO_CLOCK_GEN51_EN_DIV_DMCLK
|
374 GLAMO_CLOCK_GEN51_EN_DIV_DCLK
, 0xffff);
376 case GLAMO_ENGINE_MMC
:
377 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
378 GLAMO_HOSTBUS2_MMIO_EN_MMC
,
379 GLAMO_HOSTBUS2_MMIO_EN_MMC
);
380 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_MMC
,
381 GLAMO_CLOCK_MMC_EN_M9CLK
|
382 GLAMO_CLOCK_MMC_EN_TCLK
|
383 GLAMO_CLOCK_MMC_DG_M9CLK
|
384 GLAMO_CLOCK_MMC_DG_TCLK
, 0xffff);
385 /* enable the TCLK divider clk input */
386 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
387 GLAMO_CLOCK_GEN51_EN_DIV_TCLK
,
388 GLAMO_CLOCK_GEN51_EN_DIV_TCLK
);
390 case GLAMO_ENGINE_2D
:
391 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_2D
,
392 GLAMO_CLOCK_2D_EN_M7CLK
|
393 GLAMO_CLOCK_2D_EN_GCLK
|
394 GLAMO_CLOCK_2D_DG_M7CLK
|
395 GLAMO_CLOCK_2D_DG_GCLK
, 0xffff);
396 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
397 GLAMO_HOSTBUS2_MMIO_EN_2D
,
398 GLAMO_HOSTBUS2_MMIO_EN_2D
);
399 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
400 GLAMO_CLOCK_GEN51_EN_DIV_GCLK
,
403 case GLAMO_ENGINE_CMDQ
:
404 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_2D
,
405 GLAMO_CLOCK_2D_EN_M6CLK
, 0xffff);
406 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
407 GLAMO_HOSTBUS2_MMIO_EN_CQ
,
408 GLAMO_HOSTBUS2_MMIO_EN_CQ
);
409 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
410 GLAMO_CLOCK_GEN51_EN_DIV_MCLK
,
413 /* FIXME: Implementation */
418 glamo
->engine_enabled_bitfield
|= 1 << engine
;
423 int glamo_engine_enable(struct glamo_core
*glamo
, enum glamo_engine engine
)
427 spin_lock(&glamo
->lock
);
429 ret
= __glamo_engine_enable(glamo
, engine
);
431 spin_unlock(&glamo
->lock
);
435 EXPORT_SYMBOL_GPL(glamo_engine_enable
);
437 int __glamo_engine_disable(struct glamo_core
*glamo
, enum glamo_engine engine
)
440 case GLAMO_ENGINE_LCD
:
441 /* remove pixel clock to LCM */
442 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_LCD
,
443 GLAMO_CLOCK_LCD_EN_DCLK
, 0);
444 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_LCD
,
445 GLAMO_CLOCK_LCD_EN_DHCLK
|
446 GLAMO_CLOCK_LCD_EN_DMCLK
, 0);
447 /* kill memory clock */
448 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_LCD
,
449 GLAMO_CLOCK_LCD_EN_M5CLK
, 0);
450 /* stop dividing the clocks */
451 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
452 GLAMO_CLOCK_GEN51_EN_DIV_DHCLK
|
453 GLAMO_CLOCK_GEN51_EN_DIV_DMCLK
|
454 GLAMO_CLOCK_GEN51_EN_DIV_DCLK
, 0);
457 case GLAMO_ENGINE_MMC
:
458 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_MMC
,
459 GLAMO_CLOCK_MMC_EN_M9CLK
|
460 GLAMO_CLOCK_MMC_EN_TCLK
|
461 GLAMO_CLOCK_MMC_DG_M9CLK
|
462 GLAMO_CLOCK_MMC_DG_TCLK
, 0);
463 /* disable the TCLK divider clk input */
464 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
465 GLAMO_CLOCK_GEN51_EN_DIV_TCLK
, 0);
467 case GLAMO_ENGINE_CMDQ
:
468 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_2D
,
469 GLAMO_CLOCK_2D_EN_M6CLK
,
471 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
472 GLAMO_HOSTBUS2_MMIO_EN_CQ
,
473 GLAMO_HOSTBUS2_MMIO_EN_CQ
);
474 /* __reg_set_bit_mask(glamo, GLAMO_REG_CLOCK_GEN5_1,
475 GLAMO_CLOCK_GEN51_EN_DIV_MCLK,
478 case GLAMO_ENGINE_2D
:
479 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_2D
,
480 GLAMO_CLOCK_2D_EN_M7CLK
|
481 GLAMO_CLOCK_2D_EN_GCLK
|
482 GLAMO_CLOCK_2D_DG_M7CLK
|
483 GLAMO_CLOCK_2D_DG_GCLK
,
485 __reg_set_bit_mask(glamo
, GLAMO_REG_HOSTBUS(2),
486 GLAMO_HOSTBUS2_MMIO_EN_2D
,
487 GLAMO_HOSTBUS2_MMIO_EN_2D
);
488 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
,
489 GLAMO_CLOCK_GEN51_EN_DIV_GCLK
,
496 glamo
->engine_enabled_bitfield
&= ~(1 << engine
);
500 int glamo_engine_disable(struct glamo_core
*glamo
, enum glamo_engine engine
)
504 spin_lock(&glamo
->lock
);
506 ret
= __glamo_engine_disable(glamo
, engine
);
508 spin_unlock(&glamo
->lock
);
512 EXPORT_SYMBOL_GPL(glamo_engine_disable
);
514 static const u_int16_t engine_clock_regs
[__NUM_GLAMO_ENGINES
] = {
515 [GLAMO_ENGINE_LCD
] = GLAMO_REG_CLOCK_LCD
,
516 [GLAMO_ENGINE_MMC
] = GLAMO_REG_CLOCK_MMC
,
517 [GLAMO_ENGINE_ISP
] = GLAMO_REG_CLOCK_ISP
,
518 [GLAMO_ENGINE_JPEG
] = GLAMO_REG_CLOCK_JPEG
,
519 [GLAMO_ENGINE_3D
] = GLAMO_REG_CLOCK_3D
,
520 [GLAMO_ENGINE_2D
] = GLAMO_REG_CLOCK_2D
,
521 [GLAMO_ENGINE_MPEG_ENC
] = GLAMO_REG_CLOCK_MPEG
,
522 [GLAMO_ENGINE_MPEG_DEC
] = GLAMO_REG_CLOCK_MPEG
,
525 void glamo_engine_clkreg_set(struct glamo_core
*glamo
,
526 enum glamo_engine engine
,
527 u_int16_t mask
, u_int16_t val
)
529 reg_set_bit_mask(glamo
, engine_clock_regs
[engine
], mask
, val
);
531 EXPORT_SYMBOL_GPL(glamo_engine_clkreg_set
);
533 u_int16_t
glamo_engine_clkreg_get(struct glamo_core
*glamo
,
534 enum glamo_engine engine
)
538 spin_lock(&glamo
->lock
);
539 val
= __reg_read(glamo
, engine_clock_regs
[engine
]);
540 spin_unlock(&glamo
->lock
);
544 EXPORT_SYMBOL_GPL(glamo_engine_clkreg_get
);
546 static const struct glamo_script engine_div_regs
[__NUM_GLAMO_ENGINES
] = {
547 [GLAMO_ENGINE_LCD
] = {GLAMO_REG_CLOCK_GEN5_1
, GLAMO_CLOCK_GEN51_EN_DIV_DCLK
},
548 [GLAMO_ENGINE_MMC
] = {GLAMO_REG_CLOCK_GEN5_1
, GLAMO_CLOCK_GEN51_EN_DIV_TCLK
},
549 [GLAMO_ENGINE_2D
] = {GLAMO_REG_CLOCK_GEN5_1
, GLAMO_CLOCK_GEN51_EN_DIV_GCLK
},
552 void glamo_engine_div_enable(struct glamo_core
*glamo
, enum glamo_engine engine
)
554 uint16_t reg
= engine_div_regs
[engine
].reg
;
555 uint16_t bit
= engine_div_regs
[engine
].val
;
558 spin_lock(&glamo
->lock
);
559 val
= __reg_read(glamo
, reg
);
560 __reg_write(glamo
, reg
, val
| bit
);
561 spin_unlock(&glamo
->lock
);
564 EXPORT_SYMBOL_GPL(glamo_engine_div_enable
);
566 void glamo_engine_div_disable(struct glamo_core
*glamo
, enum glamo_engine engine
)
568 uint16_t reg
= engine_div_regs
[engine
].reg
;
569 uint16_t bit
= engine_div_regs
[engine
].val
;
572 spin_lock(&glamo
->lock
);
573 val
= __reg_read(glamo
, reg
);
574 __reg_write(glamo
, reg
, val
& ~bit
);
575 spin_unlock(&glamo
->lock
);
577 EXPORT_SYMBOL_GPL(glamo_engine_div_disable
);
579 static const struct glamo_script reset_regs
[] = {
580 [GLAMO_ENGINE_LCD
] = {
581 GLAMO_REG_CLOCK_LCD
, GLAMO_CLOCK_LCD_RESET
584 [GLAMO_ENGINE_HOST
] = {
585 GLAMO_REG_CLOCK_HOST
, GLAMO_CLOCK_HOST_RESET
587 [GLAMO_ENGINE_MEM
] = {
588 GLAMO_REG_CLOCK_MEM
, GLAMO_CLOCK_MEM_RESET
591 [GLAMO_ENGINE_MMC
] = {
592 GLAMO_REG_CLOCK_MMC
, GLAMO_CLOCK_MMC_RESET
594 [GLAMO_ENGINE_CMDQ
] = {
595 GLAMO_REG_CLOCK_2D
, GLAMO_CLOCK_2D_CQ_RESET
597 [GLAMO_ENGINE_2D
] = {
598 GLAMO_REG_CLOCK_2D
, GLAMO_CLOCK_2D_RESET
600 [GLAMO_ENGINE_JPEG
] = {
601 GLAMO_REG_CLOCK_JPEG
, GLAMO_CLOCK_JPEG_RESET
605 void glamo_engine_reset(struct glamo_core
*glamo
, enum glamo_engine engine
)
607 uint16_t reg
= reset_regs
[engine
].reg
;
608 uint16_t val
= reset_regs
[engine
].val
;
610 if (engine
>= ARRAY_SIZE(reset_regs
)) {
611 dev_warn(&glamo
->pdev
->dev
, "unknown engine %u ", engine
);
616 spin_lock(&glamo
->lock
);
617 __reg_set_bit(glamo
, reg
, val
);
618 __reg_clear_bit(glamo
, reg
, val
);
619 spin_unlock(&glamo
->lock
);
621 EXPORT_SYMBOL_GPL(glamo_engine_reset
);
623 void glamo_lcm_reset(struct platform_device
*pdev
, int level
)
625 struct glamo_core
*glamo
= dev_get_drvdata(&pdev
->dev
);
629 glamo_gpio_setpin(glamo
, GLAMO_GPIO4
, level
);
630 glamo_gpio_cfgpin(glamo
, GLAMO_GPIO4_OUTPUT
);
632 EXPORT_SYMBOL_GPL(glamo_lcm_reset
);
634 int glamo_pll_rate(struct glamo_core
*glamo
,
638 unsigned int osci
= glamo
->pdata
->osci_clock_rate
;
642 reg
= __reg_read(glamo
, GLAMO_REG_PLL_GEN1
);
645 reg
= __reg_read(glamo
, GLAMO_REG_PLL_GEN3
);
652 EXPORT_SYMBOL_GPL(glamo_pll_rate
);
654 int glamo_engine_reclock(struct glamo_core
*glamo
,
655 enum glamo_engine engine
,
659 u_int16_t reg
, mask
, div
;
665 case GLAMO_ENGINE_LCD
:
667 reg
= GLAMO_REG_CLOCK_GEN7
;
670 case GLAMO_ENGINE_MMC
:
672 reg
= GLAMO_REG_CLOCK_GEN8
;
676 dev_warn(&glamo
->pdev
->dev
,
677 "reclock of engine 0x%x not supported\n", engine
);
682 pll
= glamo_pll_rate(glamo
, pll
);
686 if (div
!= 0 && pll
/ div
<= hz
)
692 dev_dbg(&glamo
->pdev
->dev
,
693 "PLL %d, kHZ %d, div %d\n", pll
, hz
/ 1000, div
);
695 reg_set_bit_mask(glamo
, reg
, mask
, div
);
696 mdelay(5); /* wait some time to stabilize */
698 return pll
/ (div
+ 1);
700 EXPORT_SYMBOL_GPL(glamo_engine_reclock
);
702 /***********************************************************************
704 ***********************************************************************/
706 int glamo_run_script(struct glamo_core
*glamo
, const struct glamo_script
*script
,
707 int len
, int may_sleep
)
711 for (i
= 0; i
< len
; i
++) {
712 struct glamo_script
*line
= &script
[i
];
721 mdelay(line
->val
* 4);
724 /* spin until PLLs lock */
725 while ((__reg_read(glamo
, GLAMO_REG_PLL_GEN5
) & 3) != 3)
730 * couple of people reported artefacts with 2.6.28 changes, this
731 * allows reversion to 2.6.24 settings
735 switch (slow_memory
) {
736 /* choice 1 is the most conservative */
737 case 1: /* 3 waits on Async BB R & W, Use PLL 1 for mem bus */
738 __reg_write(glamo
, script
[i
].reg
, 0xef0);
740 case 2: /* 2 waits on Async BB R & W, Use PLL 1 for mem bus */
741 __reg_write(glamo
, script
[i
].reg
, 0xea0);
743 case 3: /* 1 waits on Async BB R & W, Use PLL 1 for mem bus */
744 __reg_write(glamo
, script
[i
].reg
, 0xe50);
746 case 4: /* 0 waits on Async BB R & W, Use PLL 1 for mem bus */
747 __reg_write(glamo
, script
[i
].reg
, 0xe00);
750 /* using PLL2 for memory bus increases CPU bandwidth significantly */
751 case 5: /* 3 waits on Async BB R & W, Use PLL 2 for mem bus */
752 __reg_write(glamo
, script
[i
].reg
, 0xef3);
754 case 6: /* 2 waits on Async BB R & W, Use PLL 2 for mem bus */
755 __reg_write(glamo
, script
[i
].reg
, 0xea3);
757 case 7: /* 1 waits on Async BB R & W, Use PLL 2 for mem bus */
758 __reg_write(glamo
, script
[i
].reg
, 0xe53);
760 /* default of 0 or >7 is fastest */
761 default: /* 0 waits on Async BB R & W, Use PLL 2 for mem bus */
762 __reg_write(glamo
, script
[i
].reg
, 0xe03);
768 __reg_write(glamo
, script
[i
].reg
, script
[i
].val
);
775 EXPORT_SYMBOL(glamo_run_script
);
777 static const struct glamo_script glamo_init_script
[] = {
778 { GLAMO_REG_CLOCK_HOST
, 0x1000 },
780 { GLAMO_REG_CLOCK_MEMORY
, 0x1000 },
781 { GLAMO_REG_CLOCK_MEMORY
, 0x2000 },
782 { GLAMO_REG_CLOCK_LCD
, 0x1000 },
783 { GLAMO_REG_CLOCK_MMC
, 0x1000 },
784 { GLAMO_REG_CLOCK_ISP
, 0x1000 },
785 { GLAMO_REG_CLOCK_ISP
, 0x3000 },
786 { GLAMO_REG_CLOCK_JPEG
, 0x1000 },
787 { GLAMO_REG_CLOCK_3D
, 0x1000 },
788 { GLAMO_REG_CLOCK_3D
, 0x3000 },
789 { GLAMO_REG_CLOCK_2D
, 0x1000 },
790 { GLAMO_REG_CLOCK_2D
, 0x3000 },
791 { GLAMO_REG_CLOCK_RISC1
, 0x1000 },
792 { GLAMO_REG_CLOCK_MPEG
, 0x3000 },
793 { GLAMO_REG_CLOCK_MPEG
, 0x3000 },
794 { GLAMO_REG_CLOCK_MPROC
, 0x1000 /*0x100f*/ },
796 { GLAMO_REG_CLOCK_HOST
, 0x0000 },
797 { GLAMO_REG_CLOCK_MEMORY
, 0x0000 },
798 { GLAMO_REG_CLOCK_LCD
, 0x0000 },
799 { GLAMO_REG_CLOCK_MMC
, 0x0000 },
801 /* unused engines must be left in reset to stop MMC block read "blackouts" */
802 { GLAMO_REG_CLOCK_ISP
, 0x0000 },
803 { GLAMO_REG_CLOCK_ISP
, 0x0000 },
804 { GLAMO_REG_CLOCK_JPEG
, 0x0000 },
805 { GLAMO_REG_CLOCK_3D
, 0x0000 },
806 { GLAMO_REG_CLOCK_3D
, 0x0000 },
807 { GLAMO_REG_CLOCK_2D
, 0x0000 },
808 { GLAMO_REG_CLOCK_2D
, 0x0000 },
809 { GLAMO_REG_CLOCK_RISC1
, 0x0000 },
810 { GLAMO_REG_CLOCK_MPEG
, 0x0000 },
811 { GLAMO_REG_CLOCK_MPEG
, 0x0000 },
813 { GLAMO_REG_PLL_GEN1
, 0x05db }, /* 48MHz */
814 { GLAMO_REG_PLL_GEN3
, 0x0aba }, /* 90MHz */
817 * b9 of this register MUST be zero to get any interrupts on INT#
818 * the other set bits enable all the engine interrupt sources
820 { GLAMO_REG_IRQ_ENABLE
, 0x01ff },
821 { GLAMO_REG_CLOCK_GEN6
, 0x2000 },
822 { GLAMO_REG_CLOCK_GEN7
, 0x0101 },
823 { GLAMO_REG_CLOCK_GEN8
, 0x0100 },
824 { GLAMO_REG_CLOCK_HOST
, 0x000d },
826 * b7..b4 = 0 = no wait states on read or write
827 * b0 = 1 select PLL2 for Host interface, b1 = enable it
829 { 0x200, 0x0e03 /* this is replaced by script parser */ },
835 /* S-Media recommended "set tiling mode to 512 mode for memory access
836 * more efficiency when 640x480" */
837 { GLAMO_REG_MEM_TYPE
, 0x0c74 }, /* 8MB, 16 word pg wr+rd */
838 { GLAMO_REG_MEM_GEN
, 0xafaf }, /* 63 grants min + max */
840 { GLAMO_REGOFS_HOSTBUS
+ 2, 0xffff }, /* enable on MMIO*/
842 { GLAMO_REG_MEM_TIMING1
, 0x0108 },
843 { GLAMO_REG_MEM_TIMING2
, 0x0010 }, /* Taa = 3 MCLK */
844 { GLAMO_REG_MEM_TIMING3
, 0x0000 },
845 { GLAMO_REG_MEM_TIMING4
, 0x0000 }, /* CE1# delay fall/rise */
846 { GLAMO_REG_MEM_TIMING5
, 0x0000 }, /* UB# LB# */
847 { GLAMO_REG_MEM_TIMING6
, 0x0000 }, /* OE# */
848 { GLAMO_REG_MEM_TIMING7
, 0x0000 }, /* WE# */
849 { GLAMO_REG_MEM_TIMING8
, 0x1002 }, /* MCLK delay, was 0x1000 */
850 { GLAMO_REG_MEM_TIMING9
, 0x6006 },
851 { GLAMO_REG_MEM_TIMING10
, 0x00ff },
852 { GLAMO_REG_MEM_TIMING11
, 0x0001 },
853 { GLAMO_REG_MEM_POWER1
, 0x0020 },
854 { GLAMO_REG_MEM_POWER2
, 0x0000 },
855 { GLAMO_REG_MEM_DRAM1
, 0x0000 },
857 { GLAMO_REG_MEM_DRAM1
, 0xc100 },
859 { GLAMO_REG_MEM_DRAM1
, 0xe100 },
860 { GLAMO_REG_MEM_DRAM2
, 0x01d6 },
861 { GLAMO_REG_CLOCK_MEMORY
, 0x000b },
862 { GLAMO_REG_GPIO_GEN1
, 0x000f },
863 { GLAMO_REG_GPIO_GEN2
, 0x111e },
864 { GLAMO_REG_GPIO_GEN3
, 0xccc3 },
865 { GLAMO_REG_GPIO_GEN4
, 0x111e },
866 { GLAMO_REG_GPIO_GEN5
, 0x000f },
869 static struct glamo_script glamo_resume_script
[] = {
871 { GLAMO_REG_PLL_GEN1
, 0x05db }, /* 48MHz */
872 { GLAMO_REG_PLL_GEN3
, 0x0aba }, /* 90MHz */
873 { GLAMO_REG_DFT_GEN6
, 1 },
879 * b9 of this register MUST be zero to get any interrupts on INT#
880 * the other set bits enable all the engine interrupt sources
882 { GLAMO_REG_IRQ_ENABLE
, 0x01ff },
883 { GLAMO_REG_CLOCK_HOST
, 0x0018 },
884 { GLAMO_REG_CLOCK_GEN5_1
, 0x18b1 },
886 { GLAMO_REG_MEM_DRAM1
, 0x0000 },
888 { GLAMO_REG_MEM_DRAM1
, 0xc100 },
890 { GLAMO_REG_MEM_DRAM1
, 0xe100 },
891 { GLAMO_REG_MEM_DRAM2
, 0x01d6 },
892 { GLAMO_REG_CLOCK_MEMORY
, 0x000b },
901 static void glamo_power(struct glamo_core
*glamo
,
902 enum glamo_power new_state
)
907 spin_lock_irqsave(&glamo
->lock
, flags
);
909 dev_info(&glamo
->pdev
->dev
, "***** glamo_power -> %d\n", new_state
);
913 static const REG_VALUE_MASK_TYPE reg_powerOn[] =
915 { REG_GEN_DFT6, REG_BIT_ALL, REG_DATA(1u << 0) },
916 { REG_GEN_PLL3, 0u, REG_DATA(1u << 13) },
917 { REG_GEN_MEM_CLK, REG_BIT_ALL, REG_BIT_EN_MOCACLK },
918 { REG_MEM_DRAM2, 0u, REG_BIT_EN_DEEP_POWER_DOWN },
919 { REG_MEM_DRAM1, 0u, REG_BIT_SELF_REFRESH }
922 static const REG_VALUE_MASK_TYPE reg_powerStandby[] =
924 { REG_MEM_DRAM1, REG_BIT_ALL, REG_BIT_SELF_REFRESH },
925 { REG_GEN_MEM_CLK, 0u, REG_BIT_EN_MOCACLK },
926 { REG_GEN_PLL3, REG_BIT_ALL, REG_DATA(1u << 13) },
927 { REG_GEN_DFT5, REG_BIT_ALL, REG_DATA(1u << 0) }
930 static const REG_VALUE_MASK_TYPE reg_powerSuspend[] =
932 { REG_MEM_DRAM2, REG_BIT_ALL, REG_BIT_EN_DEEP_POWER_DOWN },
933 { REG_GEN_MEM_CLK, 0u, REG_BIT_EN_MOCACLK },
934 { REG_GEN_PLL3, REG_BIT_ALL, REG_DATA(1u << 13) },
935 { REG_GEN_DFT5, REG_BIT_ALL, REG_DATA(1u << 0) }
943 * glamo state on resume is nondeterministic in some
944 * fundamental way, it has also been observed that the
945 * Glamo reset pin can get asserted by, eg, touching it with
946 * a scope probe. So the only answer is to roll with it and
947 * force an external reset on the Glamo during resume.
950 (glamo
->pdata
->glamo_external_reset
)(0);
952 (glamo
->pdata
->glamo_external_reset
)(1);
955 glamo_run_script(glamo
, glamo_init_script
,
956 ARRAY_SIZE(glamo_init_script
), 0);
960 case GLAMO_POWER_SUSPEND
:
962 /* nuke interrupts */
963 __reg_write(glamo
, GLAMO_REG_IRQ_ENABLE
, 0x200);
965 /* stash a copy of which engines were running */
966 glamo
->engine_enabled_bitfield_suspend
=
967 glamo
->engine_enabled_bitfield
;
969 /* take down each engine before we kill mem and pll */
970 for (n
= 0; n
< __NUM_GLAMO_ENGINES
; n
++)
971 if (glamo
->engine_enabled_bitfield
& (1 << n
))
972 __glamo_engine_disable(glamo
, n
);
974 /* enable self-refresh */
976 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
,
977 GLAMO_MEM_DRAM1_EN_DRAM_REFRESH
|
978 GLAMO_MEM_DRAM1_EN_GATE_CKE
|
979 GLAMO_MEM_DRAM1_SELF_REFRESH
|
980 GLAMO_MEM_REFRESH_COUNT
);
981 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
,
982 GLAMO_MEM_DRAM1_EN_MODEREG_SET
|
983 GLAMO_MEM_DRAM1_EN_DRAM_REFRESH
|
984 GLAMO_MEM_DRAM1_EN_GATE_CKE
|
985 GLAMO_MEM_DRAM1_SELF_REFRESH
|
986 GLAMO_MEM_REFRESH_COUNT
);
988 /* force RAM into deep powerdown */
990 __reg_write(glamo
, GLAMO_REG_MEM_DRAM2
,
991 GLAMO_MEM_DRAM2_DEEP_PWRDOWN
|
994 (1 << 2) | /* tRCD */
995 2); /* CAS latency */
997 /* disable clocks to memory */
998 __reg_write(glamo
, GLAMO_REG_CLOCK_MEMORY
, 0);
1000 /* all dividers from OSCI */
1001 __reg_set_bit_mask(glamo
, GLAMO_REG_CLOCK_GEN5_1
, 0x400, 0x400);
1003 /* PLL2 into bypass */
1004 __reg_set_bit_mask(glamo
, GLAMO_REG_PLL_GEN3
, 1 << 12, 1 << 12);
1006 __reg_write(glamo
, 0x200, 0x0e00);
1009 /* kill PLLS 1 then 2 */
1010 __reg_write(glamo
, GLAMO_REG_DFT_GEN5
, 0x0001);
1011 __reg_set_bit_mask(glamo
, GLAMO_REG_PLL_GEN3
, 1 << 13, 1 << 13);
1016 spin_unlock_irqrestore(&glamo
->lock
, flags
);
1020 #define MEMDETECT_RETRY 6
1021 static unsigned int detect_memsize(struct glamo_core
*glamo
)
1025 /*static const u_int16_t pattern[] = {
1026 0x1111, 0x8a8a, 0x2222, 0x7a7a,
1027 0x3333, 0x6a6a, 0x4444, 0x5a5a,
1028 0x5555, 0x4a4a, 0x6666, 0x3a3a,
1029 0x7777, 0x2a2a, 0x8888, 0x1a1a
1032 for (i
= 0; i
< MEMDETECT_RETRY
; i
++) {
1033 switch (glamo
->type
) {
1035 __reg_write(glamo
, GLAMO_REG_MEM_TYPE
, 0x0072);
1036 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
, 0xc100);
1039 switch (glamo
->revision
) {
1040 case GLAMO_CORE_REV_A0
:
1042 __reg_write(glamo
, GLAMO_REG_MEM_TYPE
,
1045 __reg_write(glamo
, GLAMO_REG_MEM_TYPE
,
1048 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
, 0x0000);
1050 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
, 0xc100);
1054 __reg_write(glamo
, GLAMO_REG_MEM_TYPE
,
1057 __reg_write(glamo
, GLAMO_REG_MEM_TYPE
,
1060 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
, 0x0000);
1062 __reg_write(glamo
, GLAMO_REG_MEM_DRAM1
, 0xe100);
1073 /* FIXME: finish implementation */
1074 for (j
= 0; j
< 8; j
++) {
1083 /* Find out if we can support this version of the Glamo chip */
1084 static int glamo_supported(struct glamo_core
*glamo
)
1086 u_int16_t dev_id
, rev_id
; /*, memsize; */
1088 dev_id
= __reg_read(glamo
, GLAMO_REG_DEVICE_ID
);
1089 rev_id
= __reg_read(glamo
, GLAMO_REG_REVISION_ID
);
1094 case GLAMO_CORE_REV_A2
:
1096 case GLAMO_CORE_REV_A0
:
1097 case GLAMO_CORE_REV_A1
:
1098 case GLAMO_CORE_REV_A3
:
1099 dev_warn(&glamo
->pdev
->dev
, "untested core revision "
1100 "%04x, your mileage may vary\n", rev_id
);
1103 dev_warn(&glamo
->pdev
->dev
, "unknown glamo revision "
1104 "%04x, your mileage may vary\n", rev_id
);
1105 /* maybe should abort ? */
1111 dev_err(&glamo
->pdev
->dev
, "unsupported Glamo device %04x\n",
1116 dev_dbg(&glamo
->pdev
->dev
, "Detected Glamo core %04x Revision %04x "
1117 "(%uHz CPU / %uHz Memory)\n", dev_id
, rev_id
,
1118 glamo_pll_rate(glamo
, GLAMO_PLL1
),
1119 glamo_pll_rate(glamo
, GLAMO_PLL2
));
1124 static int __init
glamo_probe(struct platform_device
*pdev
)
1127 struct glamo_core
*glamo
;
1129 glamo
= kmalloc(GFP_KERNEL
, sizeof(*glamo
));
1133 spin_lock_init(&glamo
->lock
);
1135 glamo
->mem
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
1136 glamo
->irq
= platform_get_irq(pdev
, 0);
1137 glamo
->pdata
= pdev
->dev
.platform_data
;
1138 if (!glamo
->mem
|| !glamo
->pdata
) {
1139 dev_err(&pdev
->dev
, "platform device with no MEM/PDATA ?\n");
1144 /* register a number of sibling devices whoise IOMEM resources
1145 * are siblings of pdev's IOMEM resource */
1147 /* only remap the generic, hostbus and memory controller registers */
1148 glamo
->base
= ioremap(glamo
->mem
->start
, 0x4000 /*GLAMO_REGOFS_VIDCAP*/);
1150 dev_err(&pdev
->dev
, "failed to ioremap() memory region\n");
1154 platform_set_drvdata(pdev
, glamo
);
1156 (glamo
->pdata
->glamo_external_reset
)(0);
1158 (glamo
->pdata
->glamo_external_reset
)(1);
1162 * finally set the mfd interrupts up
1163 * can't do them earlier or sibling probes blow up
1166 for (irq
= IRQ_GLAMO(0); irq
<= IRQ_GLAMO(8); irq
++) {
1167 set_irq_chip_and_handler(irq
, &glamo_irq_chip
, handle_level_irq
);
1168 set_irq_flags(irq
, IRQF_VALID
);
1169 set_irq_chip_data(irq
, glamo
);
1172 if (glamo
->pdata
->glamo_irq_is_wired
&&
1173 !glamo
->pdata
->glamo_irq_is_wired()) {
1174 set_irq_chained_handler(glamo
->irq
, glamo_irq_demux_handler
);
1175 set_irq_type(glamo
->irq
, IRQ_TYPE_EDGE_FALLING
);
1176 set_irq_chip_data(glamo
->irq
, glamo
);
1177 dev_info(&pdev
->dev
, "Glamo interrupt registered\n");
1178 glamo
->irq_works
= 1;
1180 dev_err(&pdev
->dev
, "Glamo interrupt not used\n");
1181 glamo
->irq_works
= 0;
1184 /* confirm it isn't insane version */
1185 if (!glamo_supported(glamo
)) {
1186 dev_err(&pdev
->dev
, "This Glamo is not supported\n");
1191 rc
= sysfs_create_group(&pdev
->dev
.kobj
, &glamo_attr_group
);
1193 dev_err(&pdev
->dev
, "cannot create sysfs group\n");
1197 /* init the chip with canned register set */
1199 dev_dbg(&glamo
->pdev
->dev
, "running init script\n");
1200 glamo_run_script(glamo
, glamo_init_script
,
1201 ARRAY_SIZE(glamo_init_script
), 1);
1203 dev_info(&glamo
->pdev
->dev
, "Glamo core PLL1: %uHz, PLL2: %uHz\n",
1204 glamo_pll_rate(glamo
, GLAMO_PLL1
),
1205 glamo_pll_rate(glamo
, GLAMO_PLL2
));
1207 /* register siblings */
1208 glamo
->pdata
->mmc_data
->core
= glamo
;
1209 glamo_cells
[GLAMO_CELL_MMC
].platform_data
= glamo
->pdata
->mmc_data
;
1210 glamo_cells
[GLAMO_CELL_MMC
].data_size
=
1211 sizeof(struct glamo_mmc_platform_data
);
1213 glamo
->pdata
->fb_data
->core
= glamo
;
1214 glamo_cells
[GLAMO_CELL_FB
].platform_data
= glamo
->pdata
->fb_data
;
1215 glamo_cells
[GLAMO_CELL_FB
].data_size
= sizeof(struct glamo_fb_platform_data
);
1217 glamo
->pdata
->spigpio_data
->core
= glamo
;
1218 glamo_cells
[GLAMO_CELL_SPI_GPIO
].platform_data
=
1219 glamo
->pdata
->spigpio_data
;
1220 glamo_cells
[GLAMO_CELL_SPI_GPIO
].data_size
=
1221 sizeof(struct glamo_spigpio_platform_data
);
1223 mfd_add_devices(&pdev
->dev
, pdev
->id
, glamo_cells
,
1224 ARRAY_SIZE(glamo_cells
),
1227 /* only request the generic, hostbus and memory controller MMIO */
1228 glamo
->mem
= request_mem_region(glamo
->mem
->start
,
1229 GLAMO_REGOFS_VIDCAP
, "glamo-core");
1231 dev_err(&pdev
->dev
, "failed to request memory region\n");
1238 disable_irq(glamo
->irq
);
1239 set_irq_chained_handler(glamo
->irq
, NULL
);
1240 set_irq_chip_data(glamo
->irq
, NULL
);
1242 for (irq
= IRQ_GLAMO(0); irq
<= IRQ_GLAMO(8); irq
++) {
1243 set_irq_flags(irq
, 0);
1244 set_irq_chip(irq
, NULL
);
1245 set_irq_chip_data(irq
, NULL
);
1248 iounmap(glamo
->base
);
1250 platform_set_drvdata(pdev
, NULL
);
1256 static int glamo_remove(struct platform_device
*pdev
)
1258 struct glamo_core
*glamo
= platform_get_drvdata(pdev
);
1261 disable_irq(glamo
->irq
);
1262 set_irq_chained_handler(glamo
->irq
, NULL
);
1263 set_irq_chip_data(glamo
->irq
, NULL
);
1265 for (irq
= IRQ_GLAMO(0); irq
<= IRQ_GLAMO(8); irq
++) {
1266 set_irq_flags(irq
, 0);
1267 set_irq_chip(irq
, NULL
);
1268 set_irq_chip_data(irq
, NULL
);
1271 platform_set_drvdata(pdev
, NULL
);
1272 mfd_remove_devices(&pdev
->dev
);
1273 iounmap(glamo
->base
);
1274 release_mem_region(glamo
->mem
->start
, GLAMO_REGOFS_VIDCAP
);
1282 static int glamo_suspend(struct device
*dev
)
1284 struct glamo_core
*glamo
= dev_get_drvdata(dev
);
1285 glamo
->suspending
= 1;
1286 glamo_power(glamo
, GLAMO_POWER_SUSPEND
);
1291 static int glamo_resume(struct device
*dev
)
1293 struct glamo_core
*glamo
= dev_get_drvdata(dev
);
1294 glamo_power(glamo
, GLAMO_POWER_ON
);
1295 glamo
->suspending
= 0;
1300 static struct dev_pm_ops glamo_pm_ops
= {
1301 .suspend
= glamo_suspend
,
1302 .resume
= glamo_resume
,
1305 #define GLAMO_PM_OPS (&glamo_pm_ops)
1308 #define GLAMO_PM_OPS NULL
1311 static struct platform_driver glamo_driver
= {
1312 .probe
= glamo_probe
,
1313 .remove
= glamo_remove
,
1315 .name
= "glamo3362",
1316 .owner
= THIS_MODULE
,
1321 static int __devinit
glamo_init(void)
1323 return platform_driver_register(&glamo_driver
);
1326 static void __exit
glamo_cleanup(void)
1328 platform_driver_unregister(&glamo_driver
);
1331 module_init(glamo_init
);
1332 module_exit(glamo_cleanup
);
1334 MODULE_AUTHOR("Harald Welte <laforge@openmoko.org>");
1335 MODULE_DESCRIPTION("Smedia Glamo 336x/337x core/resource driver");
1336 MODULE_LICENSE("GPL");