event = dm_uevent_alloc(md);
        if (!event) {
-               DMERR("%s: dm_uevent_alloc() failed", __FUNCTION__);
+               DMERR("%s: dm_uevent_alloc() failed", __func__);
                goto err_nomem;
        }
 
 
        if (add_uevent_var(&event->ku_env, "DM_TARGET=%s", ti->type->name)) {
                DMERR("%s: add_uevent_var() for DM_TARGET failed",
-                     __FUNCTION__);
+                     __func__);
                goto err_add;
        }
 
        if (add_uevent_var(&event->ku_env, "DM_ACTION=%s", dm_action)) {
                DMERR("%s: add_uevent_var() for DM_ACTION failed",
-                     __FUNCTION__);
+                     __func__);
                goto err_add;
        }
 
        if (add_uevent_var(&event->ku_env, "DM_SEQNUM=%u",
                           dm_next_uevent_seq(md))) {
                DMERR("%s: add_uevent_var() for DM_SEQNUM failed",
-                     __FUNCTION__);
+                     __func__);
                goto err_add;
        }
 
        if (add_uevent_var(&event->ku_env, "DM_PATH=%s", path)) {
-               DMERR("%s: add_uevent_var() for DM_PATH failed", __FUNCTION__);
+               DMERR("%s: add_uevent_var() for DM_PATH failed", __func__);
                goto err_add;
        }
 
        if (add_uevent_var(&event->ku_env, "DM_NR_VALID_PATHS=%d",
                           nr_valid_paths)) {
                DMERR("%s: add_uevent_var() for DM_NR_VALID_PATHS failed",
-                     __FUNCTION__);
+                     __func__);
                goto err_add;
        }
 
                if (dm_copy_name_and_uuid(event->md, event->name,
                                          event->uuid)) {
                        DMERR("%s: dm_copy_name_and_uuid() failed",
-                             __FUNCTION__);
+                             __func__);
                        goto uevent_free;
                }
 
                if (add_uevent_var(&event->ku_env, "DM_NAME=%s", event->name)) {
                        DMERR("%s: add_uevent_var() for DM_NAME failed",
-                             __FUNCTION__);
+                             __func__);
                        goto uevent_free;
                }
 
                if (add_uevent_var(&event->ku_env, "DM_UUID=%s", event->uuid)) {
                        DMERR("%s: add_uevent_var() for DM_UUID failed",
-                             __FUNCTION__);
+                             __func__);
                        goto uevent_free;
                }
 
                r = kobject_uevent_env(kobj, event->action, event->ku_env.envp);
                if (r)
-                       DMERR("%s: kobject_uevent_env failed", __FUNCTION__);
+                       DMERR("%s: kobject_uevent_env failed", __func__);
 uevent_free:
                dm_uevent_free(event);
        }
        struct dm_uevent *event;
 
        if (event_type >= ARRAY_SIZE(_dm_uevent_type_names)) {
-               DMERR("%s: Invalid event_type %d", __FUNCTION__, event_type);
+               DMERR("%s: Invalid event_type %d", __func__, event_type);
                goto out;
        }
 
 
 
                        bi->bi_bdev = rdev->bdev;
                        pr_debug("%s: for %llu schedule op %ld on disc %d\n",
-                               __FUNCTION__, (unsigned long long)sh->sector,
+                               __func__, (unsigned long long)sh->sector,
                                bi->bi_rw, i);
                        atomic_inc(&sh->count);
                        bi->bi_sector = sh->sector + rdev->data_offset;
        raid5_conf_t *conf = sh->raid_conf;
        int i;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        /* clear completed biofills */
        raid5_conf_t *conf = sh->raid_conf;
        int i;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        for (i = sh->disks; i--; ) {
        int target = sh->ops.target;
        struct r5dev *tgt = &sh->dev[target];
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        set_bit(R5_UPTODATE, &tgt->flags);
        int i;
 
        pr_debug("%s: stripe %llu block: %d\n",
-               __FUNCTION__, (unsigned long long)sh->sector, target);
+               __func__, (unsigned long long)sh->sector, target);
        BUG_ON(!test_bit(R5_Wantcompute, &tgt->flags));
 
        for (i = disks; i--; )
 {
        struct stripe_head *sh = stripe_head_ref;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        set_bit(STRIPE_OP_PREXOR, &sh->ops.complete);
        /* existing parity data subtracted */
        struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        for (i = disks; i--; ) {
         */
        int prexor = test_bit(STRIPE_OP_PREXOR, &pending);
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        for (i = disks; i--; ) {
 {
        struct stripe_head *sh = stripe_head_ref;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        set_bit(STRIPE_OP_POSTXOR, &sh->ops.complete);
        struct stripe_head *sh = stripe_head_ref;
        int disks = sh->disks, i, pd_idx = sh->pd_idx;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        for (i = disks; i--; ) {
        unsigned long flags;
        dma_async_tx_callback callback;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        /* check if prexor is active which means only process blocks
        struct stripe_head *sh = stripe_head_ref;
        int pd_idx = sh->pd_idx;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        if (test_and_clear_bit(STRIPE_OP_MOD_DMA_CHECK, &sh->ops.pending) &&
        int count = 0, pd_idx = sh->pd_idx, i;
        struct page *xor_dest = xor_srcs[count++] = sh->dev[pd_idx].page;
 
-       pr_debug("%s: stripe %llu\n", __FUNCTION__,
+       pr_debug("%s: stripe %llu\n", __func__,
                (unsigned long long)sh->sector);
 
        for (i = disks; i--; ) {
        locked++;
 
        pr_debug("%s: stripe %llu locked: %d pending: %lx\n",
-               __FUNCTION__, (unsigned long long)sh->sector,
+               __func__, (unsigned long long)sh->sector,
                locked, sh->ops.pending);
 
        return locked;