/*
 *
 * Copyright 2014 gRPC authors.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 *
 */

package grpc

import (
	
	
	
	
	
	
	
	
	
	

	
	
	
	
	
	
	
	
	iresolver 
	
	
	
	
	

	_            // To register roundrobin.
	_          // To register dns resolver.
	_  // To register passthrough resolver.
	_         // To register unix resolver.
)

const (
	// minimum time to give a connection to complete
	minConnectTimeout = 20 * time.Second
	// must match grpclbName in grpclb/grpclb.go
	grpclbName = "grpclb"
)

var (
	// ErrClientConnClosing indicates that the operation is illegal because
	// the ClientConn is closing.
	//
	// Deprecated: this error should not be relied upon by users; use the status
	// code of Canceled instead.
	ErrClientConnClosing = status.Error(codes.Canceled, "grpc: the client connection is closing")
	// errConnDrain indicates that the connection starts to be drained and does not accept any new RPCs.
	errConnDrain = errors.New("grpc: the connection is drained")
	// errConnClosing indicates that the connection is closing.
	errConnClosing = errors.New("grpc: the connection is closing")
	// invalidDefaultServiceConfigErrPrefix is used to prefix the json parsing error for the default
	// service config.
	invalidDefaultServiceConfigErrPrefix = "grpc: the provided default service config is invalid"
)

// The following errors are returned from Dial and DialContext
var (
	// errNoTransportSecurity indicates that there is no transport security
	// being set for ClientConn. Users should either set one or explicitly
	// call WithInsecure DialOption to disable security.
	errNoTransportSecurity = errors.New("grpc: no transport security set (use grpc.WithTransportCredentials(insecure.NewCredentials()) explicitly or set credentials)")
	// errTransportCredsAndBundle indicates that creds bundle is used together
	// with other individual Transport Credentials.
	errTransportCredsAndBundle = errors.New("grpc: credentials.Bundle may not be used with individual TransportCredentials")
	// errNoTransportCredsInBundle indicated that the configured creds bundle
	// returned a transport credentials which was nil.
	errNoTransportCredsInBundle = errors.New("grpc: credentials.Bundle must return non-nil transport credentials")
	// errTransportCredentialsMissing indicates that users want to transmit
	// security information (e.g., OAuth2 token) which requires secure
	// connection on an insecure connection.
	errTransportCredentialsMissing = errors.New("grpc: the credentials require transport level security (use grpc.WithTransportCredentials() to set)")
)

const (
	defaultClientMaxReceiveMessageSize = 1024 * 1024 * 4
	defaultClientMaxSendMessageSize    = math.MaxInt32
	// http2IOBufSize specifies the buffer size for sending frames.
	defaultWriteBufSize = 32 * 1024
	defaultReadBufSize  = 32 * 1024
)

// Dial creates a client connection to the given target.
func ( string,  ...DialOption) (*ClientConn, error) {
	return DialContext(context.Background(), , ...)
}

type defaultConfigSelector struct {
	sc *ServiceConfig
}

func ( *defaultConfigSelector) ( iresolver.RPCInfo) (*iresolver.RPCConfig, error) {
	return &iresolver.RPCConfig{
		Context:      .Context,
		MethodConfig: getMethodConfig(.sc, .Method),
	}, nil
}

// DialContext creates a client connection to the given target. By default, it's
// a non-blocking dial (the function won't wait for connections to be
// established, and connecting happens in the background). To make it a blocking
// dial, use WithBlock() dial option.
//
// In the non-blocking case, the ctx does not act against the connection. It
// only controls the setup steps.
//
// In the blocking case, ctx can be used to cancel or expire the pending
// connection. Once this function returns, the cancellation and expiration of
// ctx will be noop. Users should call ClientConn.Close to terminate all the
// pending operations after this function returns.
//
// The target name syntax is defined in
// https://github.com/grpc/grpc/blob/master/doc/naming.md.
// e.g. to use dns resolver, a "dns:///" prefix should be applied to the target.
func ( context.Context,  string,  ...DialOption) ( *ClientConn,  error) {
	 := &ClientConn{
		target:            ,
		csMgr:             &connectivityStateManager{},
		conns:             make(map[*addrConn]struct{}),
		dopts:             defaultDialOptions(),
		blockingpicker:    newPickerWrapper(),
		czData:            new(channelzData),
		firstResolveEvent: grpcsync.NewEvent(),
	}
	.retryThrottler.Store((*retryThrottler)(nil))
	.safeConfigSelector.UpdateConfigSelector(&defaultConfigSelector{nil})
	.ctx, .cancel = context.WithCancel(context.Background())

	for ,  := range extraDialOptions {
		.apply(&.dopts)
	}

	for ,  := range  {
		.apply(&.dopts)
	}

	chainUnaryClientInterceptors()
	chainStreamClientInterceptors()

	defer func() {
		if  != nil {
			.Close()
		}
	}()

	 := .dopts.channelzParentID
	.channelzID = channelz.RegisterChannel(&channelzChannel{}, , )
	 := &channelz.TraceEventDesc{
		Desc:     "Channel created",
		Severity: channelz.CtInfo,
	}
	if .dopts.channelzParentID != nil {
		.Parent = &channelz.TraceEventDesc{
			Desc:     fmt.Sprintf("Nested Channel(id:%d) created", .channelzID.Int()),
			Severity: channelz.CtInfo,
		}
	}
	channelz.AddTraceEvent(logger, .channelzID, 1, )
	.csMgr.channelzID = .channelzID

	if .dopts.copts.TransportCredentials == nil && .dopts.copts.CredsBundle == nil {
		return nil, errNoTransportSecurity
	}
	if .dopts.copts.TransportCredentials != nil && .dopts.copts.CredsBundle != nil {
		return nil, errTransportCredsAndBundle
	}
	if .dopts.copts.CredsBundle != nil && .dopts.copts.CredsBundle.TransportCredentials() == nil {
		return nil, errNoTransportCredsInBundle
	}
	 := .dopts.copts.TransportCredentials
	if  == nil {
		 = .dopts.copts.CredsBundle.TransportCredentials()
	}
	if .Info().SecurityProtocol == "insecure" {
		for ,  := range .dopts.copts.PerRPCCredentials {
			if .RequireTransportSecurity() {
				return nil, errTransportCredentialsMissing
			}
		}
	}

	if .dopts.defaultServiceConfigRawJSON != nil {
		 := parseServiceConfig(*.dopts.defaultServiceConfigRawJSON)
		if .Err != nil {
			return nil, fmt.Errorf("%s: %v", invalidDefaultServiceConfigErrPrefix, .Err)
		}
		.dopts.defaultServiceConfig, _ = .Config.(*ServiceConfig)
	}
	.mkp = .dopts.copts.KeepaliveParams

	if .dopts.copts.UserAgent != "" {
		.dopts.copts.UserAgent += " " + grpcUA
	} else {
		.dopts.copts.UserAgent = grpcUA
	}

	if .dopts.timeout > 0 {
		var  context.CancelFunc
		,  = context.WithTimeout(, .dopts.timeout)
		defer ()
	}
	defer func() {
		select {
		case <-.Done():
			switch {
			case .Err() == :
				 = nil
			case  == nil || !.dopts.returnLastError:
				,  = nil, .Err()
			default:
				,  = nil, fmt.Errorf("%v: %v", .Err(), )
			}
		default:
		}
	}()

	 := false
	if .dopts.scChan != nil {
		// Try to get an initial service config.
		select {
		case ,  := <-.dopts.scChan:
			if  {
				.sc = &
				.safeConfigSelector.UpdateConfigSelector(&defaultConfigSelector{&})
				 = true
			}
		default:
		}
	}
	if .dopts.bs == nil {
		.dopts.bs = backoff.DefaultExponential
	}

	// Determine the resolver to use.
	,  := .parseTargetAndFindResolver()
	if  != nil {
		return nil, 
	}
	.authority,  = determineAuthority(.parsedTarget.Endpoint(), .target, .dopts)
	if  != nil {
		return nil, 
	}
	channelz.Infof(logger, .channelzID, "Channel authority set to %q", .authority)

	if .dopts.scChan != nil && ! {
		// Blocking wait for the initial service config.
		select {
		case ,  := <-.dopts.scChan:
			if  {
				.sc = &
				.safeConfigSelector.UpdateConfigSelector(&defaultConfigSelector{&})
			}
		case <-.Done():
			return nil, .Err()
		}
	}
	if .dopts.scChan != nil {
		go .scWatcher()
	}

	var  credentials.TransportCredentials
	if  := .dopts.copts.TransportCredentials;  != nil {
		 = .Clone()
	}
	.balancerWrapper = newCCBalancerWrapper(, balancer.BuildOptions{
		DialCreds:        ,
		CredsBundle:      .dopts.copts.CredsBundle,
		Dialer:           .dopts.copts.Dialer,
		Authority:        .authority,
		CustomUserAgent:  .dopts.copts.UserAgent,
		ChannelzParentID: .channelzID,
		Target:           .parsedTarget,
	})

	// Build the resolver.
	,  := newCCResolverWrapper(, )
	if  != nil {
		return nil, fmt.Errorf("failed to build resolver: %v", )
	}
	.mu.Lock()
	.resolverWrapper = 
	.mu.Unlock()

	// A blocking dial blocks until the clientConn is ready.
	if .dopts.block {
		for {
			.Connect()
			 := .GetState()
			if  == connectivity.Ready {
				break
			} else if .dopts.copts.FailOnNonTempDialError &&  == connectivity.TransientFailure {
				if  = .connectionError();  != nil {
					,  := .(interface {
						() bool
					})
					if  && !.() {
						return nil, 
					}
				}
			}
			if !.WaitForStateChange(, ) {
				// ctx got timeout or canceled.
				if  = .connectionError();  != nil && .dopts.returnLastError {
					return nil, 
				}
				return nil, .Err()
			}
		}
	}

	return , nil
}

// chainUnaryClientInterceptors chains all unary client interceptors into one.
func ( *ClientConn) {
	 := .dopts.chainUnaryInts
	// Prepend dopts.unaryInt to the chaining interceptors if it exists, since unaryInt will
	// be executed before any other chained interceptors.
	if .dopts.unaryInt != nil {
		 = append([]UnaryClientInterceptor{.dopts.unaryInt}, ...)
	}
	var  UnaryClientInterceptor
	if len() == 0 {
		 = nil
	} else if len() == 1 {
		 = [0]
	} else {
		 = func( context.Context,  string, ,  interface{},  *ClientConn,  UnaryInvoker,  ...CallOption) error {
			return [0](, , , , , getChainUnaryInvoker(, 0, ), ...)
		}
	}
	.dopts.unaryInt = 
}

// getChainUnaryInvoker recursively generate the chained unary invoker.
func ( []UnaryClientInterceptor,  int,  UnaryInvoker) UnaryInvoker {
	if  == len()-1 {
		return 
	}
	return func( context.Context,  string, ,  interface{},  *ClientConn,  ...CallOption) error {
		return [+1](, , , , , (, +1, ), ...)
	}
}

// chainStreamClientInterceptors chains all stream client interceptors into one.
func ( *ClientConn) {
	 := .dopts.chainStreamInts
	// Prepend dopts.streamInt to the chaining interceptors if it exists, since streamInt will
	// be executed before any other chained interceptors.
	if .dopts.streamInt != nil {
		 = append([]StreamClientInterceptor{.dopts.streamInt}, ...)
	}
	var  StreamClientInterceptor
	if len() == 0 {
		 = nil
	} else if len() == 1 {
		 = [0]
	} else {
		 = func( context.Context,  *StreamDesc,  *ClientConn,  string,  Streamer,  ...CallOption) (ClientStream, error) {
			return [0](, , , , getChainStreamer(, 0, ), ...)
		}
	}
	.dopts.streamInt = 
}

// getChainStreamer recursively generate the chained client stream constructor.
func ( []StreamClientInterceptor,  int,  Streamer) Streamer {
	if  == len()-1 {
		return 
	}
	return func( context.Context,  *StreamDesc,  *ClientConn,  string,  ...CallOption) (ClientStream, error) {
		return [+1](, , , , (, +1, ), ...)
	}
}

// connectivityStateManager keeps the connectivity.State of ClientConn.
// This struct will eventually be exported so the balancers can access it.
type connectivityStateManager struct {
	mu         sync.Mutex
	state      connectivity.State
	notifyChan chan struct{}
	channelzID *channelz.Identifier
}

// updateState updates the connectivity.State of ClientConn.
// If there's a change it notifies goroutines waiting on state change to
// happen.
func ( *connectivityStateManager) ( connectivity.State) {
	.mu.Lock()
	defer .mu.Unlock()
	if .state == connectivity.Shutdown {
		return
	}
	if .state ==  {
		return
	}
	.state = 
	channelz.Infof(logger, .channelzID, "Channel Connectivity change to %v", )
	if .notifyChan != nil {
		// There are other goroutines waiting on this channel.
		close(.notifyChan)
		.notifyChan = nil
	}
}

func ( *connectivityStateManager) () connectivity.State {
	.mu.Lock()
	defer .mu.Unlock()
	return .state
}

func ( *connectivityStateManager) () <-chan struct{} {
	.mu.Lock()
	defer .mu.Unlock()
	if .notifyChan == nil {
		.notifyChan = make(chan struct{})
	}
	return .notifyChan
}

// ClientConnInterface defines the functions clients need to perform unary and
// streaming RPCs.  It is implemented by *ClientConn, and is only intended to
// be referenced by generated code.
type ClientConnInterface interface {
	// Invoke performs a unary RPC and returns after the response is received
	// into reply.
	Invoke(ctx context.Context, method string, args interface{}, reply interface{}, opts ...CallOption) error
	// NewStream begins a streaming RPC.
	NewStream(ctx context.Context, desc *StreamDesc, method string, opts ...CallOption) (ClientStream, error)
}

// Assert *ClientConn implements ClientConnInterface.
var _ ClientConnInterface = (*ClientConn)(nil)

// ClientConn represents a virtual connection to a conceptual endpoint, to
// perform RPCs.
//
// A ClientConn is free to have zero or more actual connections to the endpoint
// based on configuration, load, etc. It is also free to determine which actual
// endpoints to use and may change it every RPC, permitting client-side load
// balancing.
//
// A ClientConn encapsulates a range of functionality including name
// resolution, TCP connection establishment (with retries and backoff) and TLS
// handshakes. It also handles errors on established connections by
// re-resolving the name and reconnecting.
type ClientConn struct {
	ctx    context.Context    // Initialized using the background context at dial time.
	cancel context.CancelFunc // Cancelled on close.

	// The following are initialized at dial time, and are read-only after that.
	target          string               // User's dial target.
	parsedTarget    resolver.Target      // See parseTargetAndFindResolver().
	authority       string               // See determineAuthority().
	dopts           dialOptions          // Default and user specified dial options.
	channelzID      *channelz.Identifier // Channelz identifier for the channel.
	balancerWrapper *ccBalancerWrapper   // Uses gracefulswitch.balancer underneath.

	// The following provide their own synchronization, and therefore don't
	// require cc.mu to be held to access them.
	csMgr              *connectivityStateManager
	blockingpicker     *pickerWrapper
	safeConfigSelector iresolver.SafeConfigSelector
	czData             *channelzData
	retryThrottler     atomic.Value // Updated from service config.

	// firstResolveEvent is used to track whether the name resolver sent us at
	// least one update. RPCs block on this event.
	firstResolveEvent *grpcsync.Event

	// mu protects the following fields.
	// TODO: split mu so the same mutex isn't used for everything.
	mu              sync.RWMutex
	resolverWrapper *ccResolverWrapper         // Initialized in Dial; cleared in Close.
	sc              *ServiceConfig             // Latest service config received from the resolver.
	conns           map[*addrConn]struct{}     // Set to nil on close.
	mkp             keepalive.ClientParameters // May be updated upon receipt of a GoAway.

	lceMu               sync.Mutex // protects lastConnectionError
	lastConnectionError error
}

// WaitForStateChange waits until the connectivity.State of ClientConn changes from sourceState or
// ctx expires. A true value is returned in former case and false in latter.
//
// # Experimental
//
// Notice: This API is EXPERIMENTAL and may be changed or removed in a
// later release.
func ( *ClientConn) ( context.Context,  connectivity.State) bool {
	 := .csMgr.getNotifyChan()
	if .csMgr.getState() !=  {
		return true
	}
	select {
	case <-.Done():
		return false
	case <-:
		return true
	}
}

// GetState returns the connectivity.State of ClientConn.
//
// # Experimental
//
// Notice: This API is EXPERIMENTAL and may be changed or removed in a later
// release.
func ( *ClientConn) () connectivity.State {
	return .csMgr.getState()
}

// Connect causes all subchannels in the ClientConn to attempt to connect if
// the channel is idle.  Does not wait for the connection attempts to begin
// before returning.
//
// # Experimental
//
// Notice: This API is EXPERIMENTAL and may be changed or removed in a later
// release.
func ( *ClientConn) () {
	.balancerWrapper.exitIdle()
}

func ( *ClientConn) () {
	for {
		select {
		case ,  := <-.dopts.scChan:
			if ! {
				return
			}
			.mu.Lock()
			// TODO: load balance policy runtime change is ignored.
			// We may revisit this decision in the future.
			.sc = &
			.safeConfigSelector.UpdateConfigSelector(&defaultConfigSelector{&})
			.mu.Unlock()
		case <-.ctx.Done():
			return
		}
	}
}

// waitForResolvedAddrs blocks until the resolver has provided addresses or the
// context expires.  Returns nil unless the context expires first; otherwise
// returns a status error based on the context.
func ( *ClientConn) ( context.Context) error {
	// This is on the RPC path, so we use a fast path to avoid the
	// more-expensive "select" below after the resolver has returned once.
	if .firstResolveEvent.HasFired() {
		return nil
	}
	select {
	case <-.firstResolveEvent.Done():
		return nil
	case <-.Done():
		return status.FromContextError(.Err()).Err()
	case <-.ctx.Done():
		return ErrClientConnClosing
	}
}

var emptyServiceConfig *ServiceConfig

func () {
	 := parseServiceConfig("{}")
	if .Err != nil {
		panic(fmt.Sprintf("impossible error parsing empty service config: %v", .Err))
	}
	emptyServiceConfig = .Config.(*ServiceConfig)
}

func ( *ClientConn) ( []resolver.Address) {
	if .sc != nil {
		.applyServiceConfigAndBalancer(.sc, nil, )
		return
	}
	if .dopts.defaultServiceConfig != nil {
		.applyServiceConfigAndBalancer(.dopts.defaultServiceConfig, &defaultConfigSelector{.dopts.defaultServiceConfig}, )
	} else {
		.applyServiceConfigAndBalancer(emptyServiceConfig, &defaultConfigSelector{emptyServiceConfig}, )
	}
}

func ( *ClientConn) ( resolver.State,  error) error {
	defer .firstResolveEvent.Fire()
	.mu.Lock()
	// Check if the ClientConn is already closed. Some fields (e.g.
	// balancerWrapper) are set to nil when closing the ClientConn, and could
	// cause nil pointer panic if we don't have this check.
	if .conns == nil {
		.mu.Unlock()
		return nil
	}

	if  != nil {
		// May need to apply the initial service config in case the resolver
		// doesn't support service configs, or doesn't provide a service config
		// with the new addresses.
		.maybeApplyDefaultServiceConfig(nil)

		.balancerWrapper.resolverError()

		// No addresses are valid with err set; return early.
		.mu.Unlock()
		return balancer.ErrBadResolverState
	}

	var  error
	if .dopts.disableServiceConfig {
		channelz.Infof(logger, .channelzID, "ignoring service config from resolver (%v) and applying the default because service config is disabled", .ServiceConfig)
		.maybeApplyDefaultServiceConfig(.Addresses)
	} else if .ServiceConfig == nil {
		.maybeApplyDefaultServiceConfig(.Addresses)
		// TODO: do we need to apply a failing LB policy if there is no
		// default, per the error handling design?
	} else {
		if ,  := .ServiceConfig.Config.(*ServiceConfig); .ServiceConfig.Err == nil &&  {
			 := iresolver.GetConfigSelector()
			if  != nil {
				if len(.ServiceConfig.Config.(*ServiceConfig).Methods) != 0 {
					channelz.Infof(logger, .channelzID, "method configs in service config will be ignored due to presence of config selector")
				}
			} else {
				 = &defaultConfigSelector{}
			}
			.applyServiceConfigAndBalancer(, , .Addresses)
		} else {
			 = balancer.ErrBadResolverState
			if .sc == nil {
				// Apply the failing LB only if we haven't received valid service config
				// from the name resolver in the past.
				.applyFailingLB(.ServiceConfig)
				.mu.Unlock()
				return 
			}
		}
	}

	var  serviceconfig.LoadBalancingConfig
	if .sc != nil && .sc.lbConfig != nil {
		 = .sc.lbConfig.cfg
	}
	 := .balancerWrapper
	.mu.Unlock()

	 := .updateClientConnState(&balancer.ClientConnState{ResolverState: , BalancerConfig: })
	if  == nil {
		 =  // prefer ErrBadResolver state since any other error is
		// currently meaningless to the caller.
	}
	return 
}

// applyFailingLB is akin to configuring an LB policy on the channel which
// always fails RPCs. Here, an actual LB policy is not configured, but an always
// erroring picker is configured, which returns errors with information about
// what was invalid in the received service config. A config selector with no
// service config is configured, and the connectivity state of the channel is
// set to TransientFailure.
//
// Caller must hold cc.mu.
func ( *ClientConn) ( *serviceconfig.ParseResult) {
	var  error
	if .Err != nil {
		 = status.Errorf(codes.Unavailable, "error parsing service config: %v", .Err)
	} else {
		 = status.Errorf(codes.Unavailable, "illegal service config type: %T", .Config)
	}
	.safeConfigSelector.UpdateConfigSelector(&defaultConfigSelector{nil})
	.blockingpicker.updatePicker(base.NewErrPicker())
	.csMgr.updateState(connectivity.TransientFailure)
}

func ( *ClientConn) ( balancer.SubConn,  connectivity.State,  error) {
	.balancerWrapper.updateSubConnState(, , )
}

// newAddrConn creates an addrConn for addrs and adds it to cc.conns.
//
// Caller needs to make sure len(addrs) > 0.
func ( *ClientConn) ( []resolver.Address,  balancer.NewSubConnOptions) (*addrConn, error) {
	 := &addrConn{
		state:        connectivity.Idle,
		cc:           ,
		addrs:        ,
		scopts:       ,
		dopts:        .dopts,
		czData:       new(channelzData),
		resetBackoff: make(chan struct{}),
	}
	.ctx, .cancel = context.WithCancel(.ctx)
	// Track ac in cc. This needs to be done before any getTransport(...) is called.
	.mu.Lock()
	defer .mu.Unlock()
	if .conns == nil {
		return nil, ErrClientConnClosing
	}

	var  error
	.channelzID,  = channelz.RegisterSubChannel(, .channelzID, "")
	if  != nil {
		return nil, 
	}
	channelz.AddTraceEvent(logger, .channelzID, 0, &channelz.TraceEventDesc{
		Desc:     "Subchannel created",
		Severity: channelz.CtInfo,
		Parent: &channelz.TraceEventDesc{
			Desc:     fmt.Sprintf("Subchannel(id:%d) created", .channelzID.Int()),
			Severity: channelz.CtInfo,
		},
	})

	.conns[] = struct{}{}
	return , nil
}

// removeAddrConn removes the addrConn in the subConn from clientConn.
// It also tears down the ac with the given error.
func ( *ClientConn) ( *addrConn,  error) {
	.mu.Lock()
	if .conns == nil {
		.mu.Unlock()
		return
	}
	delete(.conns, )
	.mu.Unlock()
	.tearDown()
}

func ( *ClientConn) () *channelz.ChannelInternalMetric {
	return &channelz.ChannelInternalMetric{
		State:                    .GetState(),
		Target:                   .target,
		CallsStarted:             atomic.LoadInt64(&.czData.callsStarted),
		CallsSucceeded:           atomic.LoadInt64(&.czData.callsSucceeded),
		CallsFailed:              atomic.LoadInt64(&.czData.callsFailed),
		LastCallStartedTimestamp: time.Unix(0, atomic.LoadInt64(&.czData.lastCallStartedTime)),
	}
}

// Target returns the target string of the ClientConn.
//
// # Experimental
//
// Notice: This API is EXPERIMENTAL and may be changed or removed in a
// later release.
func ( *ClientConn) () string {
	return .target
}

func ( *ClientConn) () {
	atomic.AddInt64(&.czData.callsStarted, 1)
	atomic.StoreInt64(&.czData.lastCallStartedTime, time.Now().UnixNano())
}

func ( *ClientConn) () {
	atomic.AddInt64(&.czData.callsSucceeded, 1)
}

func ( *ClientConn) () {
	atomic.AddInt64(&.czData.callsFailed, 1)
}

// connect starts creating a transport.
// It does nothing if the ac is not IDLE.
// TODO(bar) Move this to the addrConn section.
func ( *addrConn) () error {
	.mu.Lock()
	if .state == connectivity.Shutdown {
		if logger.V(2) {
			logger.Infof("connect called on shutdown addrConn; ignoring.")
		}
		.mu.Unlock()
		return errConnClosing
	}
	if .state != connectivity.Idle {
		if logger.V(2) {
			logger.Infof("connect called on addrConn in non-idle state (%v); ignoring.", .state)
		}
		.mu.Unlock()
		return nil
	}
	// Update connectivity state within the lock to prevent subsequent or
	// concurrent calls from resetting the transport more than once.
	.updateConnectivityState(connectivity.Connecting, nil)
	.mu.Unlock()

	.resetTransport()
	return nil
}

func (,  []resolver.Address) bool {
	if len() != len() {
		return false
	}
	for ,  := range  {
		if !.Equal([]) {
			return false
		}
	}
	return true
}

// tryUpdateAddrs tries to update ac.addrs with the new addresses list.
//
// If ac is TransientFailure, it updates ac.addrs and returns true. The updated
// addresses will be picked up by retry in the next iteration after backoff.
//
// If ac is Shutdown or Idle, it updates ac.addrs and returns true.
//
// If the addresses is the same as the old list, it does nothing and returns
// true.
//
// If ac is Connecting, it returns false. The caller should tear down the ac and
// create a new one. Note that the backoff will be reset when this happens.
//
// If ac is Ready, it checks whether current connected address of ac is in the
// new addrs list.
//   - If true, it updates ac.addrs and returns true. The ac will keep using
//     the existing connection.
//   - If false, it does nothing and returns false.
func ( *addrConn) ( []resolver.Address) bool {
	.mu.Lock()
	defer .mu.Unlock()
	channelz.Infof(logger, .channelzID, "addrConn: tryUpdateAddrs curAddr: %v, addrs: %v", .curAddr, )
	if .state == connectivity.Shutdown ||
		.state == connectivity.TransientFailure ||
		.state == connectivity.Idle {
		.addrs = 
		return true
	}

	if equalAddresses(.addrs, ) {
		return true
	}

	if .state == connectivity.Connecting {
		return false
	}

	// ac.state is Ready, try to find the connected address.
	var  bool
	for ,  := range  {
		.ServerName = .cc.getServerName()
		if reflect.DeepEqual(.curAddr, ) {
			 = true
			break
		}
	}
	channelz.Infof(logger, .channelzID, "addrConn: tryUpdateAddrs curAddrFound: %v", )
	if  {
		.addrs = 
	}

	return 
}

// getServerName determines the serverName to be used in the connection
// handshake. The default value for the serverName is the authority on the
// ClientConn, which either comes from the user's dial target or through an
// authority override specified using the WithAuthority dial option. Name
// resolvers can specify a per-address override for the serverName through the
// resolver.Address.ServerName field which is used only if the WithAuthority
// dial option was not used. The rationale is that per-address authority
// overrides specified by the name resolver can represent a security risk, while
// an override specified by the user is more dependable since they probably know
// what they are doing.
func ( *ClientConn) ( resolver.Address) string {
	if .dopts.authority != "" {
		return .dopts.authority
	}
	if .ServerName != "" {
		return .ServerName
	}
	return .authority
}

func ( *ServiceConfig,  string) MethodConfig {
	if  == nil {
		return MethodConfig{}
	}
	if ,  := .Methods[];  {
		return 
	}
	 := strings.LastIndex(, "/")
	if ,  := .Methods[[:+1]];  {
		return 
	}
	return .Methods[""]
}

// GetMethodConfig gets the method config of the input method.
// If there's an exact match for input method (i.e. /service/method), we return
// the corresponding MethodConfig.
// If there isn't an exact match for the input method, we look for the service's default
// config under the service (i.e /service/) and then for the default for all services (empty string).
//
// If there is a default MethodConfig for the service, we return it.
// Otherwise, we return an empty MethodConfig.
func ( *ClientConn) ( string) MethodConfig {
	// TODO: Avoid the locking here.
	.mu.RLock()
	defer .mu.RUnlock()
	return getMethodConfig(.sc, )
}

func ( *ClientConn) () *healthCheckConfig {
	.mu.RLock()
	defer .mu.RUnlock()
	if .sc == nil {
		return nil
	}
	return .sc.healthCheckConfig
}

func ( *ClientConn) ( context.Context,  bool,  string) (transport.ClientTransport, balancer.PickResult, error) {
	return .blockingpicker.pick(, , balancer.PickInfo{
		Ctx:            ,
		FullMethodName: ,
	})
}

func ( *ClientConn) ( *ServiceConfig,  iresolver.ConfigSelector,  []resolver.Address) {
	if  == nil {
		// should never reach here.
		return
	}
	.sc = 
	if  != nil {
		.safeConfigSelector.UpdateConfigSelector()
	}

	if .sc.retryThrottling != nil {
		 := &retryThrottler{
			tokens: .sc.retryThrottling.MaxTokens,
			max:    .sc.retryThrottling.MaxTokens,
			thresh: .sc.retryThrottling.MaxTokens / 2,
			ratio:  .sc.retryThrottling.TokenRatio,
		}
		.retryThrottler.Store()
	} else {
		.retryThrottler.Store((*retryThrottler)(nil))
	}

	var  string
	if .sc != nil && .sc.lbConfig != nil {
		 = .sc.lbConfig.name
	} else {
		var  bool
		for ,  := range  {
			if .Type == resolver.GRPCLB {
				 = true
				break
			}
		}
		if  {
			 = grpclbName
		} else if .sc != nil && .sc.LB != nil {
			 = *.sc.LB
		} else {
			 = PickFirstBalancerName
		}
	}
	.balancerWrapper.switchTo()
}

func ( *ClientConn) ( resolver.ResolveNowOptions) {
	.mu.RLock()
	 := .resolverWrapper
	.mu.RUnlock()
	if  == nil {
		return
	}
	go .resolveNow()
}

// ResetConnectBackoff wakes up all subchannels in transient failure and causes
// them to attempt another connection immediately.  It also resets the backoff
// times used for subsequent attempts regardless of the current state.
//
// In general, this function should not be used.  Typical service or network
// outages result in a reasonable client reconnection strategy by default.
// However, if a previously unavailable network becomes available, this may be
// used to trigger an immediate reconnect.
//
// # Experimental
//
// Notice: This API is EXPERIMENTAL and may be changed or removed in a
// later release.
func ( *ClientConn) () {
	.mu.Lock()
	 := .conns
	.mu.Unlock()
	for  := range  {
		.resetConnectBackoff()
	}
}

// Close tears down the ClientConn and all underlying connections.
func ( *ClientConn) () error {
	defer .cancel()

	.mu.Lock()
	if .conns == nil {
		.mu.Unlock()
		return ErrClientConnClosing
	}
	 := .conns
	.conns = nil
	.csMgr.updateState(connectivity.Shutdown)

	 := .resolverWrapper
	.resolverWrapper = nil
	 := .balancerWrapper
	.mu.Unlock()

	// The order of closing matters here since the balancer wrapper assumes the
	// picker is closed before it is closed.
	.blockingpicker.close()
	if  != nil {
		.close()
	}
	if  != nil {
		.close()
	}

	for  := range  {
		.tearDown(ErrClientConnClosing)
	}
	 := &channelz.TraceEventDesc{
		Desc:     "Channel deleted",
		Severity: channelz.CtInfo,
	}
	if .dopts.channelzParentID != nil {
		.Parent = &channelz.TraceEventDesc{
			Desc:     fmt.Sprintf("Nested channel(id:%d) deleted", .channelzID.Int()),
			Severity: channelz.CtInfo,
		}
	}
	channelz.AddTraceEvent(logger, .channelzID, 0, )
	// TraceEvent needs to be called before RemoveEntry, as TraceEvent may add
	// trace reference to the entity being deleted, and thus prevent it from being
	// deleted right away.
	channelz.RemoveEntry(.channelzID)

	return nil
}

// addrConn is a network connection to a given address.
type addrConn struct {
	ctx    context.Context
	cancel context.CancelFunc

	cc     *ClientConn
	dopts  dialOptions
	acbw   balancer.SubConn
	scopts balancer.NewSubConnOptions

	// transport is set when there's a viable transport (note: ac state may not be READY as LB channel
	// health checking may require server to report healthy to set ac to READY), and is reset
	// to nil when the current transport should no longer be used to create a stream (e.g. after GoAway
	// is received, transport is closed, ac has been torn down).
	transport transport.ClientTransport // The current transport.

	mu      sync.Mutex
	curAddr resolver.Address   // The current address.
	addrs   []resolver.Address // All addresses that the resolver resolved to.

	// Use updateConnectivityState for updating addrConn's connectivity state.
	state connectivity.State

	backoffIdx   int // Needs to be stateful for resetConnectBackoff.
	resetBackoff chan struct{}

	channelzID *channelz.Identifier
	czData     *channelzData
}

// Note: this requires a lock on ac.mu.
func ( *addrConn) ( connectivity.State,  error) {
	if .state ==  {
		return
	}
	.state = 
	channelz.Infof(logger, .channelzID, "Subchannel Connectivity change to %v", )
	.cc.handleSubConnStateChange(.acbw, , )
}

// adjustParams updates parameters used to create transports upon
// receiving a GoAway.
func ( *addrConn) ( transport.GoAwayReason) {
	switch  {
	case transport.GoAwayTooManyPings:
		 := 2 * .dopts.copts.KeepaliveParams.Time
		.cc.mu.Lock()
		if  > .cc.mkp.Time {
			.cc.mkp.Time = 
		}
		.cc.mu.Unlock()
	}
}

func ( *addrConn) () {
	.mu.Lock()
	if .state == connectivity.Shutdown {
		.mu.Unlock()
		return
	}

	 := .addrs
	 := .dopts.bs.Backoff(.backoffIdx)
	// This will be the duration that dial gets to finish.
	 := minConnectTimeout
	if .dopts.minConnectTimeout != nil {
		 = .dopts.minConnectTimeout()
	}

	if  <  {
		// Give dial more time as we keep failing to connect.
		 = 
	}
	// We can potentially spend all the time trying the first address, and
	// if the server accepts the connection and then hangs, the following
	// addresses will never be tried.
	//
	// The spec doesn't mention what should be done for multiple addresses.
	// https://github.com/grpc/grpc/blob/master/doc/connection-backoff.md#proposed-backoff-algorithm
	 := time.Now().Add()

	.updateConnectivityState(connectivity.Connecting, nil)
	.mu.Unlock()

	if  := .tryAllAddrs(, );  != nil {
		.cc.resolveNow(resolver.ResolveNowOptions{})
		// After exhausting all addresses, the addrConn enters
		// TRANSIENT_FAILURE.
		.mu.Lock()
		if .state == connectivity.Shutdown {
			.mu.Unlock()
			return
		}
		.updateConnectivityState(connectivity.TransientFailure, )

		// Backoff.
		 := .resetBackoff
		.mu.Unlock()

		 := time.NewTimer()
		select {
		case <-.C:
			.mu.Lock()
			.backoffIdx++
			.mu.Unlock()
		case <-:
			.Stop()
		case <-.ctx.Done():
			.Stop()
			return
		}

		.mu.Lock()
		if .state != connectivity.Shutdown {
			.updateConnectivityState(connectivity.Idle, )
		}
		.mu.Unlock()
		return
	}
	// Success; reset backoff.
	.mu.Lock()
	.backoffIdx = 0
	.mu.Unlock()
}

// tryAllAddrs tries to creates a connection to the addresses, and stop when at
// the first successful one. It returns an error if no address was successfully
// connected, or updates ac appropriately with the new transport.
func ( *addrConn) ( []resolver.Address,  time.Time) error {
	var  error
	for ,  := range  {
		.mu.Lock()
		if .state == connectivity.Shutdown {
			.mu.Unlock()
			return errConnClosing
		}

		.cc.mu.RLock()
		.dopts.copts.KeepaliveParams = .cc.mkp
		.cc.mu.RUnlock()

		 := .dopts.copts
		if .scopts.CredsBundle != nil {
			.CredsBundle = .scopts.CredsBundle
		}
		.mu.Unlock()

		channelz.Infof(logger, .channelzID, "Subchannel picks a new address %q to connect", .Addr)

		 := .createTransport(, , )
		if  == nil {
			return nil
		}
		if  == nil {
			 = 
		}
		.cc.updateConnectionError()
	}

	// Couldn't connect to any address.
	return 
}

// createTransport creates a connection to addr. It returns an error if the
// address was not successfully connected, or updates ac appropriately with the
// new transport.
func ( *addrConn) ( resolver.Address,  transport.ConnectOptions,  time.Time) error {
	.ServerName = .cc.getServerName()
	,  := context.WithCancel(.ctx)

	 := func( transport.GoAwayReason) {
		.mu.Lock()
		defer .mu.Unlock()
		// adjust params based on GoAwayReason
		.adjustParams()
		if .state == connectivity.Shutdown {
			// Already shut down.  tearDown() already cleared the transport and
			// canceled hctx via ac.ctx, and we expected this connection to be
			// closed, so do nothing here.
			return
		}
		()
		if .transport == nil {
			// We're still connecting to this address, which could error.  Do
			// not update the connectivity state or resolve; these will happen
			// at the end of the tryAllAddrs connection loop in the event of an
			// error.
			return
		}
		.transport = nil
		// Refresh the name resolver on any connection loss.
		.cc.resolveNow(resolver.ResolveNowOptions{})
		// Always go idle and wait for the LB policy to initiate a new
		// connection attempt.
		.updateConnectivityState(connectivity.Idle, nil)
	}

	,  := context.WithDeadline(.ctx, )
	defer ()
	.ChannelzParentID = .channelzID

	,  := transport.NewClientTransport(, .cc.ctx, , , )
	if  != nil {
		if logger.V(2) {
			logger.Infof("Creating new client transport to %q: %v", , )
		}
		// newTr is either nil, or closed.
		()
		channelz.Warningf(logger, .channelzID, "grpc: addrConn.createTransport failed to connect to %s. Err: %v", , )
		return 
	}

	.mu.Lock()
	defer .mu.Unlock()
	if .state == connectivity.Shutdown {
		// This can happen if the subConn was removed while in `Connecting`
		// state. tearDown() would have set the state to `Shutdown`, but
		// would not have closed the transport since ac.transport would not
		// have been set at that point.
		//
		// We run this in a goroutine because newTr.Close() calls onClose()
		// inline, which requires locking ac.mu.
		//
		// The error we pass to Close() is immaterial since there are no open
		// streams at this point, so no trailers with error details will be sent
		// out. We just need to pass a non-nil error.
		go .Close(transport.ErrConnClosing)
		return nil
	}
	if .Err() != nil {
		// onClose was already called for this connection, but the connection
		// was successfully established first.  Consider it a success and set
		// the new state to Idle.
		.updateConnectivityState(connectivity.Idle, nil)
		return nil
	}
	.curAddr = 
	.transport = 
	.startHealthCheck() // Will set state to READY if appropriate.
	return nil
}

// startHealthCheck starts the health checking stream (RPC) to watch the health
// stats of this connection if health checking is requested and configured.
//
// LB channel health checking is enabled when all requirements below are met:
// 1. it is not disabled by the user with the WithDisableHealthCheck DialOption
// 2. internal.HealthCheckFunc is set by importing the grpc/health package
// 3. a service config with non-empty healthCheckConfig field is provided
// 4. the load balancer requests it
//
// It sets addrConn to READY if the health checking stream is not started.
//
// Caller must hold ac.mu.
func ( *addrConn) ( context.Context) {
	var  bool
	defer func() {
		if ! {
			.updateConnectivityState(connectivity.Ready, nil)
		}
	}()

	if .cc.dopts.disableHealthCheck {
		return
	}
	 := .cc.healthCheckConfig()
	if  == nil {
		return
	}
	if !.scopts.HealthCheckEnabled {
		return
	}
	 := .cc.dopts.healthCheckFunc
	if  == nil {
		// The health package is not imported to set health check function.
		//
		// TODO: add a link to the health check doc in the error message.
		channelz.Error(logger, .channelzID, "Health check is requested but health check function is not set.")
		return
	}

	 = true

	// Set up the health check helper functions.
	 := .transport
	 := func( string) (interface{}, error) {
		.mu.Lock()
		if .transport !=  {
			.mu.Unlock()
			return nil, status.Error(codes.Canceled, "the provided transport is no longer valid to use")
		}
		.mu.Unlock()
		return newNonRetryClientStream(, &StreamDesc{ServerStreams: true}, , , )
	}
	 := func( connectivity.State,  error) {
		.mu.Lock()
		defer .mu.Unlock()
		if .transport !=  {
			return
		}
		.updateConnectivityState(, )
	}
	// Start the health checking stream.
	go func() {
		 := .cc.dopts.healthCheckFunc(, , , .ServiceName)
		if  != nil {
			if status.Code() == codes.Unimplemented {
				channelz.Error(logger, .channelzID, "Subchannel health check is unimplemented at server side, thus health check is disabled")
			} else {
				channelz.Errorf(logger, .channelzID, "Health checking failed: %v", )
			}
		}
	}()
}

func ( *addrConn) () {
	.mu.Lock()
	close(.resetBackoff)
	.backoffIdx = 0
	.resetBackoff = make(chan struct{})
	.mu.Unlock()
}

// getReadyTransport returns the transport if ac's state is READY or nil if not.
func ( *addrConn) () transport.ClientTransport {
	.mu.Lock()
	defer .mu.Unlock()
	if .state == connectivity.Ready {
		return .transport
	}
	return nil
}

// tearDown starts to tear down the addrConn.
//
// Note that tearDown doesn't remove ac from ac.cc.conns, so the addrConn struct
// will leak. In most cases, call cc.removeAddrConn() instead.
func ( *addrConn) ( error) {
	.mu.Lock()
	if .state == connectivity.Shutdown {
		.mu.Unlock()
		return
	}
	 := .transport
	.transport = nil
	// We have to set the state to Shutdown before anything else to prevent races
	// between setting the state and logic that waits on context cancellation / etc.
	.updateConnectivityState(connectivity.Shutdown, nil)
	.cancel()
	.curAddr = resolver.Address{}
	if  == errConnDrain &&  != nil {
		// GracefulClose(...) may be executed multiple times when
		// i) receiving multiple GoAway frames from the server; or
		// ii) there are concurrent name resolver/Balancer triggered
		// address removal and GoAway.
		// We have to unlock and re-lock here because GracefulClose => Close => onClose, which requires locking ac.mu.
		.mu.Unlock()
		.GracefulClose()
		.mu.Lock()
	}
	channelz.AddTraceEvent(logger, .channelzID, 0, &channelz.TraceEventDesc{
		Desc:     "Subchannel deleted",
		Severity: channelz.CtInfo,
		Parent: &channelz.TraceEventDesc{
			Desc:     fmt.Sprintf("Subchannel(id:%d) deleted", .channelzID.Int()),
			Severity: channelz.CtInfo,
		},
	})
	// TraceEvent needs to be called before RemoveEntry, as TraceEvent may add
	// trace reference to the entity being deleted, and thus prevent it from
	// being deleted right away.
	channelz.RemoveEntry(.channelzID)
	.mu.Unlock()
}

func ( *addrConn) () connectivity.State {
	.mu.Lock()
	defer .mu.Unlock()
	return .state
}

func ( *addrConn) () *channelz.ChannelInternalMetric {
	.mu.Lock()
	 := .curAddr.Addr
	.mu.Unlock()
	return &channelz.ChannelInternalMetric{
		State:                    .getState(),
		Target:                   ,
		CallsStarted:             atomic.LoadInt64(&.czData.callsStarted),
		CallsSucceeded:           atomic.LoadInt64(&.czData.callsSucceeded),
		CallsFailed:              atomic.LoadInt64(&.czData.callsFailed),
		LastCallStartedTimestamp: time.Unix(0, atomic.LoadInt64(&.czData.lastCallStartedTime)),
	}
}

func ( *addrConn) () {
	atomic.AddInt64(&.czData.callsStarted, 1)
	atomic.StoreInt64(&.czData.lastCallStartedTime, time.Now().UnixNano())
}

func ( *addrConn) () {
	atomic.AddInt64(&.czData.callsSucceeded, 1)
}

func ( *addrConn) () {
	atomic.AddInt64(&.czData.callsFailed, 1)
}

type retryThrottler struct {
	max    float64
	thresh float64
	ratio  float64

	mu     sync.Mutex
	tokens float64 // TODO(dfawley): replace with atomic and remove lock.
}

// throttle subtracts a retry token from the pool and returns whether a retry
// should be throttled (disallowed) based upon the retry throttling policy in
// the service config.
func ( *retryThrottler) () bool {
	if  == nil {
		return false
	}
	.mu.Lock()
	defer .mu.Unlock()
	.tokens--
	if .tokens < 0 {
		.tokens = 0
	}
	return .tokens <= .thresh
}

func ( *retryThrottler) () {
	if  == nil {
		return
	}
	.mu.Lock()
	defer .mu.Unlock()
	.tokens += .ratio
	if .tokens > .max {
		.tokens = .max
	}
}

type channelzChannel struct {
	cc *ClientConn
}

func ( *channelzChannel) () *channelz.ChannelInternalMetric {
	return .cc.channelzMetric()
}

// ErrClientConnTimeout indicates that the ClientConn cannot establish the
// underlying connections within the specified timeout.
//
// Deprecated: This error is never returned by grpc and should not be
// referenced by users.
var ErrClientConnTimeout = errors.New("grpc: timed out when dialing")

func ( *ClientConn) ( string) resolver.Builder {
	for ,  := range .dopts.resolvers {
		if  == .Scheme() {
			return 
		}
	}
	return resolver.Get()
}

func ( *ClientConn) ( error) {
	.lceMu.Lock()
	.lastConnectionError = 
	.lceMu.Unlock()
}

func ( *ClientConn) () error {
	.lceMu.Lock()
	defer .lceMu.Unlock()
	return .lastConnectionError
}

func ( *ClientConn) () (resolver.Builder, error) {
	channelz.Infof(logger, .channelzID, "original dial target is: %q", .target)

	var  resolver.Builder
	,  := parseTarget(.target)
	if  != nil {
		channelz.Infof(logger, .channelzID, "dial target %q parse failed: %v", .target, )
	} else {
		channelz.Infof(logger, .channelzID, "parsed dial target is: %+v", )
		 = .getResolver(.URL.Scheme)
		if  != nil {
			.parsedTarget = 
			return , nil
		}
	}

	// We are here because the user's dial target did not contain a scheme or
	// specified an unregistered scheme. We should fallback to the default
	// scheme, except when a custom dialer is specified in which case, we should
	// always use passthrough scheme.
	 := resolver.GetDefaultScheme()
	channelz.Infof(logger, .channelzID, "fallback to scheme %q", )
	 :=  + ":///" + .target

	,  = parseTarget()
	if  != nil {
		channelz.Infof(logger, .channelzID, "dial target %q parse failed: %v", , )
		return nil, 
	}
	channelz.Infof(logger, .channelzID, "parsed dial target is: %+v", )
	 = .getResolver(.URL.Scheme)
	if  == nil {
		return nil, fmt.Errorf("could not get resolver for default scheme: %q", .URL.Scheme)
	}
	.parsedTarget = 
	return , nil
}

// parseTarget uses RFC 3986 semantics to parse the given target into a
// resolver.Target struct containing scheme, authority and url. Query
// params are stripped from the endpoint.
func ( string) (resolver.Target, error) {
	,  := url.Parse()
	if  != nil {
		return resolver.Target{}, 
	}

	return resolver.Target{
		Scheme:    .Scheme,
		Authority: .Host,
		URL:       *,
	}, nil
}

// Determine channel authority. The order of precedence is as follows:
// - user specified authority override using `WithAuthority` dial option
// - creds' notion of server name for the authentication handshake
// - endpoint from dial target of the form "scheme://[authority]/endpoint"
func (,  string,  dialOptions) (string, error) {
	// Historically, we had two options for users to specify the serverName or
	// authority for a channel. One was through the transport credentials
	// (either in its constructor, or through the OverrideServerName() method).
	// The other option (for cases where WithInsecure() dial option was used)
	// was to use the WithAuthority() dial option.
	//
	// A few things have changed since:
	// - `insecure` package with an implementation of the `TransportCredentials`
	//   interface for the insecure case
	// - WithAuthority() dial option support for secure credentials
	 := ""
	if  := .copts.TransportCredentials;  != nil && .Info().ServerName != "" {
		 = .Info().ServerName
	}
	 := .authority
	if ( != "" &&  != "") &&  !=  {
		return "", fmt.Errorf("ClientConn's authority from transport creds %q and dial option %q don't match", , )
	}

	switch {
	case  != "":
		return , nil
	case  != "":
		return , nil
	case strings.HasPrefix(, "unix:") || strings.HasPrefix(, "unix-abstract:"):
		// TODO: remove when the unix resolver implements optional interface to
		// return channel authority.
		return "localhost", nil
	case strings.HasPrefix(, ":"):
		return "localhost" + , nil
	default:
		// TODO: Define an optional interface on the resolver builder to return
		// the channel authority given the user's dial target. For resolvers
		// which don't implement this interface, we will use the endpoint from
		// "scheme://authority/endpoint" as the default authority.
		return , nil
	}
}