本文参考《Android系统源代码情景分析》,作者罗升阳

一、测试代码:

~/Android/external/binder/server

----FregServer.cpp

~/Android/external/binder/common

----IFregService.cpp

----IFregService.h

~/Android/external/binder/client

----FregClient.cpp


Binder库(libbinder)代码:

~/Android/frameworks/base/libs/binder

----BpBinder.cpp

----Parcel.cpp

----ProcessState.cpp

----Binder.cpp

----IInterface.cpp

----IPCThreadState.cpp

----IServiceManager.cpp

----Static.cpp

~/Android/frameworks/base/include/binder

----Binder.h

----BpBinder.h

----IInterface.h

----IPCThreadState.h

----IServiceManager.h

----IBinder.h

----Parcel.h

----ProcessState.h


驱动层代码:

~/Android//kernel/goldfish/drivers/staging/android

----binder.c

----binder.h


二、源码分析

上一篇文章Android Binder进程间通信---FregServer进程,封装进程间通信数据http://blog.csdn.net/jltxgcy/article/details/26059215,我们执行完了addService中封装数据部分代码,如下:

~/Android/frameworks/base/libs/binder

----IServiceManager.cpp

class BpServiceManager : public BpInterface<IServiceManager>{public:    .........    virtual status_t addService(const String16& name, const sp<IBinder>& service)    {        Parcel data, reply;        data.writeInterfaceToken(IServiceManager::getInterfaceDescriptor());//写入一个Binder进程间通信请求头        data.writeString16(name);//写入将要注册的Service组件的名称        data.writeStrongBinder(service);//将要注册的Service组件封装成一个flat_binder_object结构体,写入data        status_t err = remote()->transact(ADD_SERVICE_TRANSACTION, data, &reply);//remote为BpBinder对象        return err == NO_ERROR ? reply.readExceptionCode() : err;    }    ..........};
接下来调用内部的一个Binder代理对象的成员函数transact发送一个ADD_SERVICE_TRANSACTION命令协议。
remote()获取BpBinder对象,调用它的成员函数transact函数,实现如下:

~/Android/frameworks/base/libs/binder

----BpBinder.cpp

status_t BpBinder::transact(    uint32_t code, const Parcel& data, Parcel* reply, uint32_t flags)//注意data是一个引用,reply是一个指针{    // Once a binder has died, it will never come back to life.    if (mAlive) {        status_t status = IPCThreadState::self()->transact(            mHandle, code, data, reply, flags);//每个参数的介绍在下面        if (status == DEAD_OBJECT) mAlive = 0;        return status;    }    return DEAD_OBJECT;}
目前mHandle为0,code为 ADD_SERVICE_TRANSACTION,data包含了要传递给Binder驱动程序的进程间通信数据;第三个参数reply是一个输出参数,用来保存进程间通信结果,第四个参数flags用来描述这是一个同步的进程间通信请求,还是一个异步的进程间通信请求,它是一个默认参数,默认值为0,表示这是一个同步的进程请求。

调用上文http://blog.csdn.net/jltxgcy/article/details/25953361已经创建的IPCThreadState对象的成员函数transact,实现如下:

~/Android/frameworks/base/libs/binder

----IPCThreadState.cpp

status_t IPCThreadState::transact(int32_t handle,                                  uint32_t code, const Parcel& data,                                  Parcel* reply, uint32_t flags){    status_t err = data.errorCheck();    flags |= TF_ACCEPT_FDS;//flags等于0|TF_ACCEPT_FDS    .......    if (err == NO_ERROR) {        ..........        err = writeTransactionData(BC_TRANSACTION, flags, handle, code, data, NULL);//将data的内容写入到一个binder_transaction_data结构体中    }    ........    if ((flags & TF_ONE_WAY) == 0) {//最后一位为0,按位与TF_ONE_WAY等于0,表示同步的进程间通信请求        ..........        if (reply) {//reply不为空,是一个指针            err = waitForResponse(reply);        } else {            .........        }        ..........    } else {        ..........    }        return err;}
其中, TF_ACCEPT_FDS,TF_ONE_WAY定义在一个枚举中。实现如下:

~/Android/frameworks/base/include/binder

----binder.h

enum transaction_flags { TF_ONE_WAY = 0x01, TF_ROOT_OBJECT = 0x04, TF_STATUS_CODE = 0x08, TF_ACCEPT_FDS = 0x10,};

首先检查Parcel对象中的进程间通信数据是否正确,然后将参数flags的TF_ACCEPT_FDS位设置为1,表示允许Server进程在返回结果中携带文件描述符。如果Parcel对象data中的进程间通信数据没有问题,那么就会调用成员函数writeTransactionData将它的内容写入到一个binder_transaction_data结构体中。接着判断flags的TF_ONE_WAY位是否等于0。如果是,那么就说明这是一个同步的进程间通信请求,这时候如果用来保存通信结果的Parcel对象reply不等于NULL,那么就调用成员函数waitForResponse函数。

我们先来分析writeTransactionData函数,实现如下:

~/Android/frameworks/base/include/binder
----IPCThreadState.cpp

status_t IPCThreadState::writeTransactionData(int32_t cmd, uint32_t binderFlags,//此时cmd为BC_TRANSACTION,binderFlags为0|TF_ACCEPT_FDS    int32_t handle, uint32_t code, const Parcel& data, status_t* statusBuffer)//handle为0,code为ADD_SERVICE_TRANSACTION,data包含了包含了要传递给Binder驱动程序的进程间通信数据,statusBuffer为NULL{    binder_transaction_data tr;    tr.target.handle = handle;//0    tr.code = code;//ADD_SERVICE_TRANSACTION    tr.flags = binderFlags;//0|TF_ACCEPT_FDS        const status_t err = data.errorCheck();    if (err == NO_ERROR) {        tr.data_size = data.ipcDataSize();//数据缓冲区大小        tr.data.ptr.buffer = data.ipcData();//数据缓冲区的起始位置        tr.offsets_size = data.ipcObjectsCount()*sizeof(size_t);//偏移数组大小        tr.data.ptr.offsets = data.ipcObjects();//偏移数组起始位置    } else if (statusBuffer) {        ........    } else {        ........    }        mOut.writeInt32(cmd);//BC_TRANSACTION    mOut.write(&tr, sizeof(tr));        return NO_ERROR;}
其中binder_transaction_data结构体,实现如下:

struct binder_transaction_data { union {    size_t handle;    void *ptr; } target; void *cookie; unsigned int code; unsigned int flags; pid_t sender_pid; uid_t sender_euid; size_t data_size; size_t offsets_size; union {    struct {       const void *buffer;       const void *offsets;    }ptr;    uint8_t buf[8]; } data;};
执行完writeTransactionData,此时命令协议缓冲区mOut的内存布局如下图:

FregServer进程,发送BC_TRANSACTION,唤醒ServiceManager进程,返回BR_TRANSACTION_COMPLETE,睡眠等待主线程thread->wait_第1张图片
注意此图有错误,flat_binder_object中cookie为BBinder类指针,即binder本地对象指针。binder为本地对象弱引用计数的地址。


执行完writeTransactionData函数,该执行waitForResponse函数了,实现如下:

~/Android/frameworks/base/include/binder

----IPCThreadState.cpp

status_t IPCThreadState::waitForResponse(Parcel *reply, status_t *acquireResult){    int32_t cmd;    int32_t err;    while (1) {        if ((err=talkWithDriver()) < NO_ERROR) break;        .....    }    .......}
这个函数通过一个while循环不断地调用成员函数talkWithDriver来与Binder驱动程序进行交互,以便可以将前面准备好的BC_TRANSACTION命令协议发送给Binder驱动程序处理,并等待Binder驱动程序将进程间通信结果返回来。

talkWithDriver函数实现如下:

~/Android/frameworks/base/include/binder

----IPCThreadState.cpp

status_t IPCThreadState::talkWithDriver(bool doReceive){    LOG_ASSERT(mProcess->mDriverFD >= 0, "Binder driver is not opened");        binder_write_read bwr;        // Is the read buffer empty?    const bool needRead = mIn.dataPosition() >= mIn.dataSize();//needRead为1,表示有需要读的数据        // We don't want to write anything if we are still reading    // from data left in the input buffer and the caller    // has requested to read the next data.    const size_t outAvail = (!doReceive || needRead) ? mOut.dataSize() : 0;//doReceive为true表示,只接受Binder驱动程序发送给该进程的返回协议。默认为true。doReceive为false,即不只接受Binder驱动程序发送给该进程的返回协议,或者有需要读的数据,那么outAvail就不为0        bwr.write_size = outAvail;//要写入的数据大小,为mOut的实际大小    bwr.write_buffer = (long unsigned int)mOut.data();//要写入的数据起始位置    // This is what we'll read.    if (doReceive && needRead) {        bwr.read_size = mIn.dataCapacity();//读入数据的大小,为它的所有容量        bwr.read_buffer = (long unsigned int)mIn.data();//读入数据的起始位置    } else {        bwr.read_size = 0;    }    .........    // Return immediately if there is nothing to do.    if ((bwr.write_size == 0) && (bwr.read_size == 0)) return NO_ERROR;//如果两者都为0,就不用继续执行驱动程序了        bwr.write_consumed = 0;//消费清0    bwr.read_consumed = 0;//消费清0    status_t err;    do {        .........#if defined(HAVE_ANDROID_OS)        if (ioctl(mProcess->mDriverFD, BINDER_WRITE_READ, &bwr) >= 0)//IPCThreadState在构造函数初始化时,初始化了mProcess            err = NO_ERROR;        else            err = -errno;#else        err = INVALID_OPERATION;#endif        IF_LOG_COMMANDS() {            alog << "Finished read/write, write size = " << mOut.dataSize() << endl;        }    } while (err == -EINTR);    ........    if (err >= NO_ERROR) {        if (bwr.write_consumed > 0) {            if (bwr.write_consumed < (ssize_t)mOut.dataSize())                mOut.remove(0, bwr.write_consumed);            else                mOut.setDataSize(0);        }        if (bwr.read_consumed > 0) {            mIn.setDataSize(bwr.read_consumed);            mIn.setDataPosition(0);        }        .........        return NO_ERROR;    }        return err;}
在IPCThreadState类内部,除了使用缓冲区mOut来保存即将要发送给Binder驱动程序的命令协议外,还使用缓冲区mIn来保存那些从Binder驱动程序接受到的返回协议。

通过IPCThreadState类的mIn,mOut设置一个局部变量binder_write_read结构体,最后调用了ioctl映射到Binder驱动程序去执行。

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

static long binder_ioctl(struct file *filp, unsigned int cmd, unsigned long arg){int ret;struct binder_proc *proc = filp->private_data;struct binder_thread *thread;unsigned int size = _IOC_SIZE(cmd);void __user *ubuf = (void __user *)arg;//上面传递的过来的局部变量binder_write_read结构体的地址.........mutex_lock(&binder_lock);thread = binder_get_thread(proc);//由于在初始化Process过程中,调用了open_driver,在open_driver中也调用了ioctl,所以looper为0if (thread == NULL) {ret = -ENOMEM;goto err;}switch (cmd) {//cmd为上面传递过来的BINDER_WRITE_READcase BINDER_WRITE_READ: {struct binder_write_read bwr;if (size != sizeof(struct binder_write_read)) {ret = -EINVAL;goto err;}if (copy_from_user(&bwr, ubuf, sizeof(bwr))) {//从用户空间传进来的一个binder_write_read结构体拷贝出来,并且保存在变量bwr中,里面有数据地址,数据总大小,数据现在消费了多少ret = -EFAULT;goto err;}                .........if (bwr.write_size > 0) {//bwr.write_size大于0,执行这里ret = binder_thread_write(proc, thread, (void __user *)bwr.write_buffer, bwr.write_size, &bwr.write_consumed);if (ret < 0) {bwr.read_consumed = 0;if (copy_to_user(ubuf, &bwr, sizeof(bwr)))ret = -EFAULT;goto err;}}if (bwr.read_size > 0) {//bwr.read_size大于0,执行这里ret = binder_thread_read(proc, thread, (void __user *)bwr.read_buffer, bwr.read_size, &bwr.read_consumed, filp->f_flags & O_NONBLOCK);if (!list_empty(&proc->todo))wake_up_interruptible(&proc->wait);if (ret < 0) {if (copy_to_user(ubuf, &bwr, sizeof(bwr)))ret = -EFAULT;goto err;}}                ...........if (copy_to_user(ubuf, &bwr, sizeof(bwr))) {//将结果返回用户空间bwrret = -EFAULT;goto err;}break;}..........ret = 0;err:if (thread)thread->looper &= ~BINDER_LOOPER_STATE_NEED_RETURN;//looper还是0mutex_unlock(&binder_lock);        ...........return ret;}

由于write_size大于0,read_size大于0,所以首先会执行binder_thread_write,然后执行binder_thread_read。

由于IPCThreadState::self()->incWeakHandle(handle),mOut中除了刚才写入的内容,还有增加弱引用计数,增加强引用计数。所以binder_thread_write首先会执行BC_INCREFS和BC_ACQUIRE命令协议,参考http://blog.csdn.net/jltxgcy/article/details/27638717,会为FregServer进程的代理对象(句柄值为0),创建一个FregServer的引用对象(node为Service Manager实体对象,proc为FregServer进程)。执行结果为FregServer的引用对象强引用计数为1,弱引用计数为1,Service Manager实体对象外部强指针为1。

binder_thread_write函数实现如下:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

intbinder_thread_write(struct binder_proc *proc, struct binder_thread *thread,    void __user *buffer, int size, signed long *consumed)//注意consumed为指针{uint32_t cmd;void __user *ptr = buffer + *consumed;//起始位置void __user *end = buffer + size;//末尾位置while (ptr < end && thread->return_error == BR_OK) {if (get_user(cmd, (uint32_t __user *)ptr))//cmd为BC_TRANSACTIONreturn -EFAULT;ptr += sizeof(uint32_t);//取出命令后,ptr自增长......switch (cmd) {......case BC_TRANSACTION:case BC_REPLY: {struct binder_transaction_data tr;if (copy_from_user(&tr, ptr, sizeof(tr)))//将进程间通信数据读取出来,并且保存在binder_transation_data结构体tr中return -EFAULT;ptr += sizeof(tr);binder_transaction(proc, thread, &tr, cmd == BC_REPLY);//调用函数binder_transaction来处理进程发送给它的BC_TRANSACTION命令协议break;}                .......*consumed = ptr - buffer;//consumed为传入的参数size,因此数据已经全部使用完毕}return 0;}
从前面的调用过程可以知道,这个输出缓冲区包含了一个BC_TRANSACTION命令协议,接着就将BC_TRANSACTION命令协议后面所跟的进程间通信数据读取出来,并且保存在binder_transation_data结构体tr中。最后调用函数binder_transaction来处理进程发送给它的BC_TRANSACTION命令协议。

函数binder_transaction负责处理命令协议BC_TRANSACTION和命令BC_REPLY,它的实现比较长,我们来分段阅读:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

static voidbinder_transaction(struct binder_proc *proc, struct binder_thread *thread,struct binder_transaction_data *tr, int reply){struct binder_transaction *t;struct binder_work *tcomplete;size_t *offp, *off_end;struct binder_proc *target_proc;struct binder_thread *target_thread = NULL;struct binder_node *target_node = NULL;struct list_head *target_list;wait_queue_head_t *target_wait;struct binder_transaction *in_reply_to = NULL;        .........if (reply) {//为0.........} else {if (tr->target.handle) {//为0struct binder_ref *ref;ref = binder_get_ref(proc, tr->target.handle);.......target_node = ref->node;} else {target_node = binder_context_mgr_node;//Service Manager的Binder实体对象binder_context_mgr_node........}........target_proc = target_node->proc;//找到了目标进程........if (!(tr->flags & TF_ONE_WAY) && thread->transaction_stack) {//目前thread->transaction_stack为NULL,所以暂时不执行这里struct binder_transaction *tmp;tmp = thread->transaction_stack;........while (tmp) {if (tmp->from && tmp->from->proc == target_proc)target_thread = tmp->from;tmp = tmp->from_parent;}}}if (target_thread) {//target_thread为NULL........target_list = &target_thread->todo;target_wait = &target_thread->wait;} else {target_list = &target_proc->todo;//target_list和target_wait分别指向该目标进程target_proc的todo队列和wait等待队列target_wait = &target_proc->wait;}
找到了target_node,target_proc,初始化了target_list和target_wait。

函数继续执行:

        t = kzalloc(sizeof(*t), GFP_KERNEL);//分配了一个binder_transaction结构体t        ......tcomplete = kzalloc(sizeof(*tcomplete), GFP_KERNEL);//分配了一个binder_work结构体tcomplete......if (!reply && !(tr->flags & TF_ONE_WAY))//如果正在处理BC_TRANSATION命令协议,而且是一个同步的进程间通信请求t->from = thread;//from指向源线程thread,以便目标进程target_proc或者目标线程target_thread处理完该进程间通信请求之后,能够找回发送该进程间通信请求的线程,最终将进程间通信结果返回给它。elset->from = NULL;t->sender_euid = proc->tsk->cred->euid;t->to_proc = target_proc;//目标进程t->to_thread = target_thread;//目标线程t->code = tr->code;//ADD_SERVICE_TRANCATIONt->flags = tr->flags;//TF_ACCEPTS_FDSt->priority = task_nice(current);t->buffer = binder_alloc_buf(target_proc, tr->data_size,tr->offsets_size, !reply && (t->flags & TF_ONE_WAY));//分配了binder_buffer结构体.......t->buffer->allow_user_free = 0;//不允许释放.......t->buffer->transaction = t;t->buffer->target_node = target_node;//binder_context_mgr_nodeif (target_node)binder_inc_node(target_node, 1, 0, NULL);//增加目标Binder实体对象的内部强引用计数offp = (size_t *)(t->buffer->data + ALIGN(tr->data_size, sizeof(void *)));//偏移数组在data中起始位置,位于数据缓冲区之后if (copy_from_user(t->buffer->data, tr->data.ptr.buffer, tr->data_size)) {//数据缓冲区拷贝到data中.......goto err_copy_data_failed;}if (copy_from_user(offp, tr->data.ptr.offsets, tr->offsets_size)) {//偏移数组拷贝到data中,偏移数组位于数据缓冲区之后.......goto err_copy_data_failed;}........off_end = (void *)offp + tr->offsets_size;//偏移数组在data中的结束位置

binder_inc_node(target_node, 1, 0, NULL); 增加目标Service Manager实体对象的内部强引用计数,目前为2。

binder_transaction结构体实现如下:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

struct binder_transaction {int debug_id;struct binder_work work;struct binder_thread *from;struct binder_transaction *from_parent;struct binder_proc *to_proc;struct binder_thread *to_thread;struct binder_transaction *to_parent;unsigned need_reply : 1;/*unsigned is_dead : 1;*/ /* not used at the moment */struct binder_buffer *buffer;unsigned intcode;unsigned intflags;longpriority;longsaved_priority;uid_tsender_euid;};
其中,binder_buffer结构体实现如下:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

struct binder_buffer {        struct list_head entry; /* free and allocated entries by addesss */        struct rb_node rb_node; /* free entry by size or allocated entry */                                /* by address */        unsigned free : 1;        unsigned allow_user_free : 1;        unsigned async_transaction : 1;        unsigned debug_id : 29;        struct binder_transaction *transaction;        struct binder_node *target_node;        size_t data_size;        size_t offsets_size;        uint8_t data[0];};

首先分配了一个binder_transaction结构体t,然后初始化各参数。

函数binder_transaction接着执行:

   for (; offp < off_end; offp++) {struct flat_binder_object *fp;        .......fp = (struct flat_binder_object *)(t->buffer->data + *offp);//取得flat_binder_object结构体switch (fp->type) {//BINDER_TYPE_BINDERcase BINDER_TYPE_BINDER:case BINDER_TYPE_WEAK_BINDER: {struct binder_ref *ref;struct binder_node *node = binder_get_node(proc, fp->binder);//第一次取得为NULLif (node == NULL) {//执行这里node = binder_new_node(proc, fp->binder, fp->cookie);//创建FregServer进程实体对象if (node == NULL) {return_error = BR_FAILED_REPLY;goto err_binder_new_node_failed;}node->min_priority = fp->flags & FLAT_BINDER_FLAG_PRIORITY_MASK;node->accept_fds = !!(fp->flags & FLAT_BINDER_FLAG_ACCEPTS_FDS);//accept_fds为1}........ref = binder_get_ref_for_node(target_proc, node);//创建Service Manager引用对象(node为FregServer进程实体对象,进程为Service Manager本地对象)........if (fp->type == BINDER_TYPE_BINDER)fp->type = BINDER_TYPE_HANDLE;//type变成了HANDLEelsefp->type = BINDER_TYPE_WEAK_HANDLE;fp->handle = ref->desc;//由于上面type变成了HANDLE,所以这里也设置handle,句柄值为1binder_inc_ref(ref, fp->type == BINDER_TYPE_HANDLE, &thread->todo);//增加引用计数.......} break;.......}

由于调用函数binder_get_node无法获得一个引用了它的Binder实体对象,所以调用函数binder_new_node为它创建一个Binder实体对象node。binder_new_node实现如下:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

static struct binder_node *binder_new_node(struct binder_proc *proc, void __user *ptr, void __user *cookie){struct rb_node **p = &proc->nodes.rb_node;struct rb_node *parent = NULL;struct binder_node *node;while (*p) {//根据node的ptr,来查找是否已经分配了binder_node结构体parent = *p;node = rb_entry(parent, struct binder_node, rb_node);if (ptr < node->ptr)p = &(*p)->rb_left;else if (ptr > node->ptr)p = &(*p)->rb_right;elsereturn NULL;}node = kzalloc(sizeof(*node), GFP_KERNEL);//如果没有找到,分配binder_node结构体if (node == NULL)return NULL;binder_stats.obj_created[BINDER_STAT_NODE]++;rb_link_node(&node->rb_node, parent, p);//根据ptr将node->rb_node插入proc->nodes中rb_insert_color(&node->rb_node, &proc->nodes);node->debug_id = ++binder_last_id;//初始化各个变量node->proc = proc;node->ptr = ptr;//弱引用计数地址node->cookie = cookie;//BBinder(Binder本地对象地址)node->work.type = BINDER_WORK_NODE;INIT_LIST_HEAD(&node->work.entry);INIT_LIST_HEAD(&node->async_todo);if (binder_debug_mask & BINDER_DEBUG_INTERNAL_REFS)printk(KERN_INFO "binder: %d:%d node %d u%p c%p created\n",       proc->pid, current->pid, node->debug_id,       node->ptr, node->cookie);return node;}
返回for循环,接下来执行,binder_get_ref_for_node,实现如下:

~/Android/kernel/goldfish/drivers/staging/android

----binder.c

static struct binder_ref *binder_get_ref_for_node(struct binder_proc *proc, struct binder_node *node){struct rb_node *n;struct rb_node **p = &proc->refs_by_node.rb_node;struct rb_node *parent = NULL;struct binder_ref *ref, *new_ref;while (*p) {//根据node,在proc->refs_by_node中来查找是否已经分配了binder_ref结构体parent = *p;ref = rb_entry(parent, struct binder_ref, rb_node_node);if (node < ref->node)p = &(*p)->rb_left;else if (node > ref->node)p = &(*p)->rb_right;elsereturn ref;}new_ref = kzalloc(sizeof(*ref), GFP_KERNEL);//如果没有找到,分配binder_ref结构体if (new_ref == NULL)return NULL;binder_stats.obj_created[BINDER_STAT_REF]++;        ........new_ref->proc = proc;new_ref->node = node;//刚刚创建的noderb_link_node(&new_ref->rb_node_node, parent, p);////将new_ref->rb_node_node插入proc->refs_by_node中rb_insert_color(&new_ref->rb_node_node, &proc->refs_by_node);new_ref->desc = (node == binder_context_mgr_node) ? 0 : 1;//为1for (n = rb_first(&proc->refs_by_desc); n != NULL; n = rb_next(n)) {ref = rb_entry(n, struct binder_ref, rb_node_desc);if (ref->desc > new_ref->desc)break;new_ref->desc = ref->desc + 1;}p = &proc->refs_by_desc.rb_node;while (*p) {parent = *p;ref = rb_entry(parent, struct binder_ref, rb_node_desc);if (new_ref->desc < ref->desc)p = &(*p)->rb_left;else if (new_ref->desc > ref->desc)p = &(*p)->rb_right;elseBUG();}rb_link_node(&new_ref->rb_node_desc, parent, p);//将new_ref->rb_node_desc插入到proc->refs_by_desc中rb_insert_color(&new_ref->rb_node_desc, &proc->refs_by_desc);if (node) {hlist_add_head(&new_ref->node_entry, &node->refs);......} else {......}return new_ref;}

获取了Binder引用对象后,重新设置了flat_binder_object结构体的type为BINDER_TYPE_HANDLE(原来为BINDER_TYPE_BINDER),所以设置fp->handle(原来为fp->binder)为refs->decs。

binder_inc_ref(ref, fp->type == BINDER_TYPE_HANDLE, &thread->todo),执行此句代码,Service Manager进程引用对象强引用为1,FregServer进程实体对象外部强引用计数为1。并且加入一个工作项来通知该进程增加FregServer进程的本地对象的强引用计数。等到下次使用IO控制命令BINDER_WRITE_READ进入到Binder驱动程序时,会执行。

函数接着执行:

       if (reply) {.....} else if (!(t->flags & TF_ONE_WAY)) {BUG_ON(t->buffer->async_transaction != 0);t->need_reply = 1;//need_reply为1t->from_parent = thread->transaction_stack;//t->from_parent为NULLthread->transaction_stack = t;//FregServer进程主线程的transaction为t} else {BUG_ON(target_node == NULL);BUG_ON(t->buffer->async_transaction != 1);if (target_node->has_async_transaction) {target_list = &target_node->async_todo;target_wait = NULL;} elsetarget_node->has_async_transaction = 1;}t->work.type = BINDER_WORK_TRANSACTION;list_add_tail(&t->work.entry, target_list);//加入到目标进程的todotcomplete->type = BINDER_WORK_TRANSACTION_COMPLETE;list_add_tail(&tcomplete->entry, &thread->todo);//加入到本线程的todoif (target_wait)wake_up_interruptible(target_wait);//唤醒目标进程return;
我们假设本线程继续执行,执行完毕后再执行被唤醒的目标进程。

返回到binder_ioctl,继续执行binder_thread_read,实现如下:

static intbinder_thread_read(struct binder_proc *proc, struct binder_thread *thread,void  __user *buffer, int size, signed long *consumed, int non_block){void __user *ptr = buffer + *consumed;void __user *end = buffer + size;int ret = 0;int wait_for_proc_work;if (*consumed == 0) {if (put_user(BR_NOOP, (uint32_t __user *)ptr))return -EFAULT;ptr += sizeof(uint32_t);}retry:wait_for_proc_work = thread->transaction_stack == NULL && list_empty(&thread->todo);//wait_for_proc_work为0        .......thread->looper |= BINDER_LOOPER_STATE_WAITING;//looper为BINDER_LOOPER_STATE_WAITINGif (wait_for_proc_work)proc->ready_threads++;mutex_unlock(&binder_lock);if (wait_for_proc_work) {........} else {if (non_block) {if (!binder_has_thread_work(thread))ret = -EAGAIN;} else//执行这里ret = wait_event_interruptible(thread->wait, binder_has_thread_work(thread));//由于thread上有数据,那么继续执行,不会睡眠等待}mutex_lock(&binder_lock);if (wait_for_proc_work)proc->ready_threads--;thread->looper &= ~BINDER_LOOPER_STATE_WAITING;//looper为0if (ret)return ret;while (1) {uint32_t cmd;struct binder_transaction_data tr;struct binder_work *w;struct binder_transaction *t = NULL;if (!list_empty(&thread->todo))w = list_first_entry(&thread->todo, struct binder_work, entry);//将线程thread的todo队列中类型为BINDER_WORK_TRANSACTION_COMPLETE的工作项取出来else if (!list_empty(&proc->todo) && wait_for_proc_work)w = list_first_entry(&proc->todo, struct binder_work, entry);else {if (ptr - buffer == 4 && !(thread->looper & BINDER_LOOPER_STATE_NEED_RETURN)) /* no data added */goto retry;break;}........switch (w->type) {//刚取出来,类型为BINDER_WORK_TRANSACTION_COMPLETE.......case BINDER_WORK_TRANSACTION_COMPLETE: {cmd = BR_TRANSACTION_COMPLETE;if (put_user(cmd, (uint32_t __user *)ptr))//将一个BR_TRANSACTION_COMPLETE返回协议写入到用户提供的缓冲区。return -EFAULT;ptr += sizeof(uint32_t);.......list_del(&w->entry);//删除todo上的工作项kfree(w);//释放结构体.......} break;........*consumed = ptr - buffer;//消耗的大小...........return 0;}

首先,将线程thread的todo队列中类型为BINDER_WORK_TRANSACTION_COMPLETE的工作项取出来,将一个BR_TRANSACTION_COMPLETE返回协议写入到用户提供的缓冲区。

除了上面的工作项需要处理,还有BINDER_WORK_NODE需要处理,增加FregServer进程实体对象的内部强引用计数和弱引用计数,都为1。并且将BR_INCREFS和BR_ACQUIRE命令返回返回协议写入到用户提供的缓冲区。

函数binder_thread_read执行完成之后,就返回到函数binder_ioctl中,然后再返回到IPCThreadState类的成员函数talkWithDriver中,在此函数中,还有一段代码要执行:

talkWithDriver函数实现如下:

~/Android/frameworks/base/include/binder

----IPCThreadState.cpp

   if (err >= NO_ERROR) {        if (bwr.write_consumed > 0) {            if (bwr.write_consumed < (ssize_t)mOut.dataSize())                mOut.remove(0, bwr.write_consumed);//mOut中移除已经写过的数据            else                mOut.setDataSize(0);        }        if (bwr.read_consumed > 0) {            mIn.setDataSize(bwr.read_consumed);//mIn中设置了读取数据的大小,即BR_TRANSACTION_COMPLETE大小            mIn.setDataPosition(0);//mIn中设置了读取数据的位置,为0,指向了BR_TRANSACTION_COMPLETE开头        }        .........        return NO_ERROR;    }        return err;
最后,又返回到IPCThreadState类的成员函数waitForResponse,实现如下:

~/Android/frameworks/base/include/binder

----IPCThreadState.cpp

status_t IPCThreadState::waitForResponse(Parcel *reply, status_t *acquireResult){    int32_t cmd;    int32_t err;    while (1) {        if ((err=talkWithDriver()) < NO_ERROR) break;        err = mIn.errorCheck();        if (err < NO_ERROR) break;        if (mIn.dataAvail() == 0) continue;                cmd = mIn.readInt32();//读取了BR_TRANSACTION_COMPLETE协议,mIn中就没有数据了        .....        switch (cmd) {        case BR_TRANSACTION_COMPLETE:            if (!reply && !acquireResult) goto finish;//reply不为NULL            break;        .....    }    ........    return err;}

继续往下执行,读取的cmd为BR_TRANSACTION_COMPLETE,执行switch,break跳出switch函数,继续执行while循环,又开始执行talkWithDriver,由于IPCThreadState类内部的命令协议缓冲区mOut中的命令协议,以及返回协议缓冲区mIn中的返回协议都已经处理完成了,

继续执行while循环,由于读取的数据还没有读完,所以talkWithDriver直接返回,参考http://blog.csdn.net/jltxgcy/article/details/27638717一文,增加FregServer进程本地对象的强引用计数和弱引用计数。都为1。

命令都执行完了,mout中没有数据,mIn中也没有数据,所以当它再次通过IO控制命令BINDER_WRITE_READ进入到Binder驱动程序binder_ioctl中时,就会调用函数binder_thread_read睡眠等待目标进程将上次发出的进程间通信请求的结果返回来。

由于最后thread->transaction_stack != NULL,所以睡眠在这句代码上。

ret = wait_event_interruptible(thread->wait, binder_has_thread_work(thread));

三、目前为止各对象的引用计数

Service Manager进程的本地对象,不需要引用计数,因为永远不会销毁。

Service Manager进程的实体对象,内部强引用计数为2,内部弱引用计数为1,外部强指针为1,has_strong_refs为1,has_week_refs为1。

FregServer进程的引用对象(引用了Service Manager进程的实体对象),强引用计数为1,弱引用计数为1。

FregServer进程的代理对象,强引用计数为1,弱引用计数为1。


FregServer进程的本地对象,强引用计数为1,弱引用计数为1。

FregServer进程的实体对象外部强引用计数为1,内部强引用计数为1,内部弱引用计数为1。

Service Manager进程的引用对象(引用了FregServer进程的实体对象)强引用为1。

更多相关文章

  1. Android 进程内存、CPU使用查看
  2. Android 中调试手段 打印函数调用栈信息
  3. Android 任务、进程和线程
  4. Android 进程回收之LowMemoryKiller原理篇
  5. Android 利用JNI调用Android Java代码函数
  6. Android中SensorManager.getRotationMatrix函数原理解释
  7. Android -- SharedPreferences保存基本数据、序列化对象、List数

随机推荐

  1. Android图形系统草稿
  2. Android版本变化而产生的权限变更
  3. Android(安卓)一个自定义View需要实现哪
  4. Android嵌套滑动机制源码分析
  5. Android(安卓)多渠道打包进阶版
  6. [转]Android(安卓)开源项目分类汇总
  7. android 用代码编写linearlayout布局
  8. Android开发小技巧(二)-------如何获取全局
  9. Watchdog 工作机制解析
  10. 自定义按钮实现android 返回按钮 事件