Changeset - 68719ba74601
[Not reviewed]
default
0 2 1
Laman - 6 years ago 2019-03-22 16:27:34

position detection, refactored out color quantization
3 files changed with 119 insertions and 53 deletions:
0 comments (0 inline, 0 general)
exp/board_detect.py
Show inline comments
 
import sys
 

	
 
sys.path.append("../src")
 

	
 
import os
 
import random
 
import itertools
 
import logging as log
 

	
 
import cv2 as cv
 
import numpy as np
 
import scipy.cluster
 
import scipy.ndimage
 
import scipy.signal
 

	
 
from geometry import Line
 
from ransac import DiagonalRansac
 
from quantization import kmeans,QuantizedImage
 
from annotations import DataFile,computeBoundingBox
 
from hough import show,prepareEdgeImg,HoughTransform
 
from analyzer.epoint import EPoint
 
from analyzer.corners import Corners
 

	
 
random.seed(361)
 
log.basicConfig(level=log.DEBUG,format="%(message)s")
 

	
 

	
 
def kmeans(img):
 
	arr=np.reshape(img,(-1,3)).astype(np.float)
 
	wood=[193,165,116]
 
	(centers,distortion)=scipy.cluster.vq.kmeans(arr,3)
 
	log.debug("k-means centers: %s",centers)
 
	(black,empty,white)=sorted(centers,key=sum)
 
	if np.linalg.norm(black)>np.linalg.norm(black-wood):
 
		black=None
 
	if np.linalg.norm(white-[255,255,255])>np.linalg.norm(white-wood):
 
		white=None
 
	log.debug("black, white: %s, %s",black,white)
 
	return (black,white,centers)
 

	
 

	
 
def quantize(img,centers):
 
	origShape=img.shape
 
	data=np.reshape(img,(-1,3))
 
	(keys,dists)=scipy.cluster.vq.vq(data,centers)
 
	pixels=np.array([centers[k] for k in keys],dtype=np.uint8).reshape(origShape)
 
	return pixels
 

	
 

	
 
def filterStones(contours,bwImg,stoneDims):
 
	contourImg=cv.cvtColor(bwImg,cv.COLOR_GRAY2BGR)
 
	res=[]
 
	for (i,c) in enumerate(contours):
 
		keep=True
 
		moments=cv.moments(c)
 
		center=(moments["m10"]/(moments["m00"] or 1), moments["m01"]/(moments["m00"] or 1))
 
		area=cv.contourArea(c)
 
		(x,y,w,h)=cv.boundingRect(c)
 
		if w>stoneDims[0] or h>stoneDims[1]*1.5 or w<2 or h<2:
 
			cv.drawMarker(contourImg,tuple(map(int,center)),(0,0,255),cv.MARKER_TILTED_CROSS,12)
 
			keep=False
 
		coverage1=area/(w*h or 1)
 
		hull=cv.convexHull(c)
 
		coverage2=area/(cv.contourArea(hull) or 1)
 
		# if coverage2<0.8:
 
		# 	cv.drawMarker(contourImg,tuple(map(int,center)),(0,127,255),cv.MARKER_DIAMOND,12)
 
		# 	keep=False
 
		if keep:
 
			res.append((EPoint(*center),c))
 
			cv.drawMarker(contourImg,tuple(map(int,center)),(255,0,0),cv.MARKER_CROSS)
 
	log.debug("accepted: %s",len(res))
 
	log.debug("rejected: %s",len(contours)-len(res))
 
	show(contourImg,"accepted and rejected stones")
 
	return res
 

	
 

	
 
class BoardDetector:
 
	def __init__(self,annotationsPath):
 
		self._annotations=DataFile(annotationsPath)
 

	
 
		self._rectW=0
 
		self._rectH=0
 
		self._rect=None
 

	
 
		self._hough=None
 
		self._rectiMatrix=None
 
		self._inverseMatrix=None
 

	
 
		self.grid=None
 

	
 
	def __call__(self,img,filename):
 
		# approximately detect the board
 
		(h,w)=img.shape[:2]
 
		log.debug("image dimensions: %s x %s",w,h)
 
		show(img,filename)
 
		(x1,y1,x2,y2)=self._detectRough(img,filename)
 
		rect=img[y1:y2,x1:x2]
 
		self._rectW=x2-x1
 
		self._rectH=y2-y1
 
		self._rect=rect
 

	
 
		# quantize colors
 
		(black,white,colors)=self._sampleColors(rect)
 
		quantized=quantize(rect,colors)
 
		quantized=QuantizedImage(rect)
 
		gray=cv.cvtColor(rect,cv.COLOR_BGR2GRAY)
 
		edges=cv.Canny(gray,70,130)
 
		show(edges,"edges")
 
		quantized=quantized & (255-cv.cvtColor(edges,cv.COLOR_GRAY2BGR))
 
		show(quantized,"quantized, edges separated")
 
		edgeMask=(255-edges)
 
		quantizedImg=quantized.img & cv.cvtColor(edgeMask,cv.COLOR_GRAY2BGR)
 
		show(quantizedImg,"quantized, edges separated")
 

	
 
		# detect black and white stones
 
		stones=self._detectStones(quantized,black,white)
 
		stones=self._detectStones(quantized,edgeMask)
 

	
 
		# detect lines from edges and stones
 
		edgeImg=prepareEdgeImg(rect)
 
		self._hough=HoughTransform(edgeImg)
 
		stonesImg=np.zeros((self._rectH,self._rectW),np.uint8)
 
		for (point,c) in stones:
 
			cv.circle(stonesImg,(int(point.x),int(point.y)),2,255,-1)
 

	
 
		show(stonesImg,"detected stones")
 
		self._hough.update(stonesImg,10)
 
		lines=self._hough.extract()
 

	
 
		linesImg=np.copy(rect)
 
		for line in itertools.chain(*lines):
 
			self._drawLine(linesImg,line)
 
		show(linesImg,"detected lines")
 

	
 
		# # rectify the image
 
		# rectify the image
 
		matrix=self._computeTransformationMatrix(lines[0][0],lines[0][-1],lines[1][0],lines[1][-1])
 
		transformed=cv.warpPerspective(rect,matrix,(self._rectW,self._rectH))
 
		rectiLines=[[line.transform(matrix) for line in pack] for pack in lines]
 
		quantized.transform(matrix)
 

	
 
		# determine precise board edges
 
		self._detectBestGrid(rectiLines,linesImg)
 
		self.grid=self._detectGrid(rectiLines,linesImg)
 

	
 
		self.detectPosition(quantized)
 

	
 
	def _detectRough(self,img,filename):
 
		corners=self._annotations[filename][0]
 
		(x1,y1,x2,y2)=computeBoundingBox(corners)
 
		log.debug("bounding box: (%s,%s) - (%s,%s)",x1,y1,x2,y2)
 
		return (x1,y1,x2,y2)
 

	
 
	def _sampleColors(self,rect):
 
		(h,w)=rect.shape[:2]
 
		minirect=rect[h//4:3*h//4, w//4:3*w//4]
 
		return kmeans(minirect)
 

	
 
	def _detectStones(self,quantized,black,white):
 
		(h,w)=quantized.shape[:2]
 
		mask=self._maskStones(quantized,black,white)
 
	def _detectStones(self,quantized,edgeMask):
 
		(h,w)=quantized.img.shape[:2]
 
		mask=self._maskStones(quantized,edgeMask)
 
		stoneDims=(w/19,h/19)
 
		log.debug("stone dims: %s - %s",tuple(x/2 for x in stoneDims),stoneDims)
 

	
 
		(contours,hierarchy)=cv.findContours(mask,cv.RETR_LIST,cv.CHAIN_APPROX_SIMPLE)
 
		stoneLocs=filterStones(contours,mask,stoneDims)
 

	
 
		return stoneLocs
 

	
 
	def _maskStones(self,quantized,black,white):
 
		unit=np.array([1,1,1],dtype=np.uint8)
 
		if black is not None:
 
			maskB=cv.inRange(quantized,black-unit,black+unit)
 
	def _maskStones(self,quantized,edgeMask):
 
		distTransform=cv.distanceTransform(quantized.maskB&edgeMask,cv.DIST_L2,5)
 
		maskB=cv.inRange(distTransform,6,20)
 
		show(maskB,"black areas")
 

	
 
			distTransform=cv.distanceTransform(maskB,cv.DIST_L2,5)
 
			maskB=cv.inRange(distTransform,6,20)
 
			show(maskB,"black areas")
 
		else: maskB=np.zeros(quantized.shape[:2],dtype=np.uint8)
 

	
 
		if white is not None:
 
			maskW=cv.inRange(quantized,white-unit,white+unit)
 
			distTransform=cv.distanceTransform(maskW,cv.DIST_L2,5)
 
			maskW=cv.inRange(distTransform,6,20)
 
			show(maskW,"white areas")
 
		else: maskW=np.zeros(quantized.shape[:2],dtype=np.uint8)
 
		distTransform=cv.distanceTransform(quantized.maskW&edgeMask,cv.DIST_L2,5)
 
		maskW=cv.inRange(distTransform,6,20)
 
		show(maskW,"white areas")
 

	
 
		stones=cv.bitwise_or(maskB,maskW)
 
		show(stones,"black and white areas")
 
		return stones
 

	
 
	def _computeTransformationMatrix(self,p,q,r,s): # p || q, r || s
 
		(a,b,c,d)=Corners([p.intersect(r),p.intersect(s),q.intersect(r),q.intersect(s)]) # canonize the abcd order
 
		pad=20
 
		a_=EPoint(b.x+pad,min(a.y,d.y)+pad)
 
		b_=EPoint(b.x+pad,max(b.y,c.y)-pad)
 
		c_=EPoint(c.x-pad,max(b.y,c.y)-pad)
 
		d_=EPoint(c.x-pad,min(a.y,d.y)+pad)
 
		abcd=[list(point) for point in (a,b,c,d)]
 
		abcd_=[list(point) for point in (a_,b_,c_,d_)]
 
		log.debug("abcd: %s ->",(a,b,c,d))
 
		log.debug("-> abcd_: %s",(a_,b_,c_,d_))
 
		matrix=cv.getPerspectiveTransform(np.float32(abcd),np.float32(abcd_))
 
		log.debug("transformation matrix: %s",matrix)
 

	
 
		rect=np.copy(self._rect)
 
		for point in (a,b,c,d):
 
			cv.drawMarker(rect,(int(point.x),int(point.y)),(0,255,255),cv.MARKER_TILTED_CROSS)
 
		show(rect)
 
		transformed=cv.warpPerspective(rect,matrix,(self._rectW,self._rectH))
 
		show(transformed,"rectified image")
 

	
 
		self._rectiMatrix=matrix
 
		self._inverseMatrix=np.linalg.inv(matrix)
 
		return matrix
 

	
 
	def _detectBestGrid(self,lines,img):
 
	def _detectGrid(self,lines,img):
 
		intersections=[]
 
		for p in lines[0]:
 
			for q in lines[1]:
 
				intersections.append(p.intersect(q))
 

	
 
		sack=DiagonalRansac(intersections,19)
 
		diagonals=sack.extract(10,3000)
 
		log.debug("diagonals candidates: %s",diagonals)
 
		for line in diagonals:
 
			self._drawLine(img,line.transform(self._inverseMatrix),[0,255,255])
 
		show(img,"diagonals candidates")
 

	
 
		best=(0,None)
 
		transformedImg=cv.warpPerspective(img,self._rectiMatrix,(self._rectW,self._rectH))
 
		explored=[0,0,0]
 

	
 
		for e in diagonals:
 
			for f in diagonals:
 
				explored[0]+=1
 
				center=e.intersect(f)
 
				if not center: continue
 
				if center.x<0 or center.x>self._rectW or center.y<0 or center.y>self._rectH: continue
 
				for line in itertools.chain(*lines):
 
					for i in range(1,10): # 10th is useless, 11-19 are symmetrical to 1-9
 
						explored[1]+=1
 
						grid=self._constructGrid(e,f,line,i)
 
						if not grid: continue
 
						explored[2]+=1
 
						score=self._scoreGrid(grid)
 
						if score>best[0]:
 
							best=(score,grid)
 
							log.debug("new best grid: %s",score)
 
							self._showGrid(transformedImg,grid)
 
		log.debug("diagonal pairs: %s, explored grids: %s, scored grids: %s",*explored)
 
		return best[1]
 

	
 
	def _constructGrid(self,e,f,line,i):
 
		"""Contruct a grid.
 

	
 
		:param e: (Line) one diagonal
 
		:param f: (Line) other diagonal
 
		:param line: (Line) one of the grid lines
 
		:param i: (int) line's index among the grid's lines, 1<=i<=9"""
 
		center=e.intersect(f)
 
		p1=line.intersect(e)
 
		p2=line.intersect(f)
 
		a=center+9*(p1-center)/(10-i)
 
		b=center+9*(p2-center)/(10-i)
 
		c=2*center-a
 
		d=2*center-b
 
		# abort unfitting sizes
 
		if not all(0<=point.x<self._rectW and 0<=point.y<self._rectH for point in (a,b,c,d)):
 
			return None
 
		if any(g.dist(h)<19*10 for (g,h) in [(a,b),(a,c),(a,d),(b,c),(b,d),(c,d)]):
 
			return None
 
		(a,b,c,d)=Corners([a,b,c,d])
 
		rows=[]
 
		cols=[]
 
		for j in range(19):
 
			rows.append(Line.fromPoints((a*(18-j)+b*j)/18,(d*(18-j)+c*j)/18))
 
			cols.append(Line.fromPoints((a*(18-j)+d*j)/18,(b*(18-j)+c*j)/18))
 
		return (rows,cols)
 

	
 
	def _scoreGrid(self,lines):
 
		(p,q,r,s)=(lines[0][0],lines[0][-1],lines[-1][0],lines[-1][-1])
 
		corners=(p.intersect(r),p.intersect(s),q.intersect(r),q.intersect(s))
 
		origCorners=[c.transform(self._inverseMatrix) for c in corners]
 
		# must fit
 
		if not all(0<=c.x<self._rectW and 0<=c.y<self._rectH for c in origCorners):
 
			return 0
 
		return sum(self._hough.scoreLine(p.transform(self._inverseMatrix)) for p in itertools.chain(*lines))
 

	
 
	def detectPosition(self,img):
 
		(rows,cols)=self.grid
 
		intersections=[[row.intersect(col) for col in cols] for row in rows]
 
		position=[[self._detectStoneAt(img,point) for point in row] for row in intersections]
 
		log.debug("detected position:\n%s","\n".join("".join(row) for row in position))
 
		return position
 

	
 
	def _detectStoneAt(self,img,intersection):
 
		(height,width)=img.img.shape[:2]
 
		(x,y)=map(int,intersection)
 
		scores=[0,0,0]
 
		for xi in range(x-2,x+3):
 
			if xi<0 or xi>=width: continue
 
			for yi in range(y-2,y+3):
 
				if yi<0 or yi>=height: continue
 
				scores[img.get(xi,yi)]+=1
 
		return sorted(list(zip(scores,"XO.")))[-1][1]
 

	
 
	def _drawLine(self,img,line,color=None):
 
		if not color: color=[0,255,0]
 
		(h,w)=img.shape[:2]
 
		corners=[EPoint(0,0),EPoint(w,0),EPoint(0,h),EPoint(w,h)] # NW NE SW SE
 
		borders=[
 
			[Line.fromPoints(corners[0],corners[1]), Line.fromPoints(corners[2],corners[3])], # N S
 
			[Line.fromPoints(corners[0],corners[2]), Line.fromPoints(corners[1],corners[3])] # W E
 
		]
 

	
 
		(a,b)=(line.intersect(borders[0][0]), line.intersect(borders[0][1]))
 
		log.debug("%s %s",line,(a,b))
 
		if not a or not b:
 
			(a,b)=(line.intersect(borders[1][0]), line.intersect(borders[1][1]))
 
			log.debug("* %s %s",line,(a,b))
 
		if any(abs(x)>10**5 for x in [*a,*b]):
 
			log.debug("ignored")
 
			return
 
		cv.line(img,(int(a.x),int(a.y)),(int(b.x),int(b.y)),color)
 

	
 
	def _showGrid(self,img,lines):
 
		img=np.copy(img)
 
		(rows,cols)=lines
 
		for row in rows:
 
			for col in cols:
exp/hough.py
Show inline comments
 
@@ -77,49 +77,49 @@ class HoughTransform:
 
		for (r,row) in enumerate(img):
 
			for (c,pix) in enumerate(row):
 
				if pix==0: continue
 
				for alphaDeg in range(0,180):
 
					d=self._computeDist(c,r,alphaDeg)+self._diagLen//2
 
					self._acc[(alphaDeg,d)]+=weight
 
		log.debug("Hough updated in %s s",round(datetime.now().timestamp()-start,3))
 

	
 
	def scoreLine(self,line):
 
		transformed=self._transformInput(line)
 
		alphaDeg=round(transformed.alpha*180/math.pi)%180
 
		d=round(transformed.d+self._diagLen//2)
 
		if not 0<=d<self._diagLen: return 0
 
		return self._acc[(alphaDeg,d)]
 

	
 
	def show(self,img=None):
 
		if img is None: img=self._createImg()
 
		show(img,"Hough transform accumulator")
 

	
 
	def _computeDist(self,x,y,alphaDeg):
 
		alphaRad=alphaDeg*math.pi/180
 
		(x0,y0)=self._center
 
		(dx,dy)=(x-x0,y0-y)
 
		d=dx*math.cos(alphaRad)+dy*math.sin(alphaRad)
 
		return round(d)
 
		return int(d)
 

	
 
	def _detectLines(self):
 
		bag=LineBag()
 
		for alpha in range(0,180+60,2):
 
			for beta in range(max(alpha-60,0),min(alpha+60,180+60),2):
 
				accLine=[self._acc[key] for key in self._readLineKeys(alpha,beta)]
 
				(peaks,props)=scipy.signal.find_peaks(accLine,prominence=0)
 
				(prominences,peaks)=zip(*sorted(zip(props["prominences"],peaks),reverse=True)[:19])
 
				bag.put(sum(prominences),alpha,beta,peaks)
 
		return bag.pull(2)
 

	
 
	def _readLineKeys(self,alpha,beta):
 
		n=self._diagLen-1
 
		res=[]
 
		for i in range(n+1):
 
			k=round((alpha*(n-i)+beta*i)/n)
 
			if k>=180:
 
				k=k%180
 
				i=n-i
 
			res.append((k,i))
 
		return res
 

	
 
	def _transformInput(self,line):
 
		reflectedLine=Line(math.pi*2-line.alpha,line.d)
exp/quantization.py
Show inline comments
 
new file 100644
 
import logging as log
 

	
 
import numpy as np
 
import scipy.cluster
 
import cv2 as cv
 

	
 

	
 
def kmeans(img):
 
	arr=np.reshape(img,(-1,3)).astype(np.float)
 
	wood=[193,165,116]
 
	(centers,distortion)=scipy.cluster.vq.kmeans(arr,3)
 
	log.debug("k-means centers: %s",centers)
 
	(black,empty,white)=sorted(centers,key=sum)
 
	if np.linalg.norm(black)>np.linalg.norm(black-wood):
 
		black=None
 
	if np.linalg.norm(white-[255,255,255])>np.linalg.norm(white-wood):
 
		white=None
 
	log.debug("black, white: %s, %s",black,white)
 
	return (black,white,centers)
 

	
 

	
 
class QuantizedImage:
 
	BLACK=0
 
	WHITE=1
 
	EMPTY=2
 

	
 
	def __init__(self,img):
 
		self.img=self._quantize(img)
 
		self._mask()
 

	
 
	def transform(self,matrix):
 
		(h,w)=self.img.shape[:2]
 
		self.img=cv.warpPerspective(self.img,matrix,(w,h))
 
		self.maskB=cv.warpPerspective(self.maskB,matrix,(w,h))
 
		self.maskW=cv.warpPerspective(self.maskW,matrix,(w,h))
 

	
 
	def get(self,x,y):
 
		if self.maskB[y,x]: return self.BLACK
 
		elif self.maskW[y,x]: return self.WHITE
 
		else: return self.EMPTY
 

	
 
	def _quantize(self,img):
 
		(self._black,self._white,colors)=self._sampleColors(img)
 
		origShape=img.shape
 
		data=np.reshape(img,(-1,3))
 
		(keys,dists)=scipy.cluster.vq.vq(data,colors)
 
		pixels=np.array([colors[k] for k in keys],dtype=np.uint8).reshape(origShape)
 
		return pixels
 

	
 
	def _sampleColors(self,rect):
 
		(h,w)=rect.shape[:2]
 
		minirect=rect[h//4:3*h//4, w//4:3*w//4]
 
		return kmeans(minirect)
 

	
 
	def _mask(self):
 
		unit=np.array([1,1,1],dtype=np.uint8)
 
		if self._black is not None:
 
			self.maskB=cv.inRange(self.img,self._black-unit,self._black+unit)
 
		else:
 
			self.maskB=np.zeros(self.img.shape[:2],dtype=np.uint8)
 

	
 
		if self._white is not None:
 
			self.maskW=cv.inRange(self.img,self._white-unit,self._white+unit)
 
		else:
 
			self.maskW=np.zeros(self.img.shape[:2],dtype=np.uint8)
0 comments (0 inline, 0 general)