}
}
-void *__snd_kmalloc(size_t size, int flags, void *caller)
+static void *__snd_kmalloc(size_t size, int flags, void *caller)
{
unsigned long cpu_flags;
struct snd_alloc_track *t;
{
unsigned long flags;
struct snd_alloc_track *t;
- if (obj == NULL) {
- snd_printk(KERN_WARNING "null kfree (called from %p)\n", __builtin_return_address(0));
+ if (obj == NULL)
return;
- }
t = snd_alloc_track_entry(obj);
if (t->magic != KMALLOC_MAGIC) {
snd_printk(KERN_WARNING "bad kfree (called from %p)\n", __builtin_return_address(0));
void snd_hidden_vfree(void *obj)
{
struct snd_alloc_track *t;
- if (obj == NULL) {
- snd_printk(KERN_WARNING "null vfree (called from %p)\n", __builtin_return_address(0));
+ if (obj == NULL)
return;
- }
t = snd_alloc_track_entry(obj);
if (t->magic != VMALLOC_MAGIC) {
snd_printk(KERN_ERR "bad vfree (called from %p)\n", __builtin_return_address(0));
*
* Returns zero if successful, or non-zero on failure.
*/
-int copy_to_user_fromio(void __user *dst, unsigned long src, size_t count)
+int copy_to_user_fromio(void __user *dst, const volatile void __iomem *src, size_t count)
{
#if defined(__i386__) || defined(CONFIG_SPARC32)
return copy_to_user(dst, (const void*)src, count) ? -EFAULT : 0;
size_t c = count;
if (c > sizeof(buf))
c = sizeof(buf);
- memcpy_fromio(buf, (void*)src, c);
+ memcpy_fromio(buf, (void __iomem *)src, c);
if (copy_to_user(dst, buf, c))
return -EFAULT;
count -= c;
*
* Returns zero if successful, or non-zero on failure.
*/
-int copy_from_user_toio(unsigned long dst, const void __user *src, size_t count)
+int copy_from_user_toio(volatile void __iomem *dst, const void __user *src, size_t count)
{
#if defined(__i386__) || defined(CONFIG_SPARC32)
return copy_from_user((void*)dst, src, count) ? -EFAULT : 0;
c = sizeof(buf);
if (copy_from_user(buf, src, c))
return -EFAULT;
- memcpy_toio((void*)dst, buf, c);
+ memcpy_toio(dst, buf, c);
count -= c;
dst += c;
src += c;