throbber
Espacenet- Bibliographic data
`
`Page | of 1
`
` Espacenet
`
`Bibliographic data: JP2O71S5733078 (A} — 2018-07-23
`
`OBJECT RECOGNITION DEVICE FOR CONTROL, MOBILE APPARATUS
`CONTROL SYSTEM, AND PROGRAM FOR OBJECT RECOGNITION FOR
`CONTROL.
`
`Inventor(s):
`
`Applicant(s):
`
`Classification:
`
`Application number:
`
`GUANHAIKE + (68 3853)
`
`RICOH CO LTD + (#kxlestt Y 4 —)
`
`-international:
`- cooperative:
`JP20140005417 20140115
`
`G067T7/00; GO6T?7/20; GO8G1/16
`
`Global Dossier
`
`Priority number(s):
`
`JP20140005417 20140115
`
`Also published as:
`
`JP6315308 (B2)
`
`Abstract of JP2078133078 (4}
`
`PROBLEM TO BE SOLVED: To achieve the appropriate control of predetermined
`equioment mounied on a mobile body by appropriately identifying whether or not an
`object existing in the periphery of the mobile body is an obiect for controh SOLUTION:
`An abject recognition part 205 detects a detection ablect such as another vehicie ora
`pedestrian existing in ihe perighery of a mobile body on the hasis of a pickup image
`obtained by picking up the periphery of the mobile body. On the other hand, an object
`state determination part 207 specifies relative positions Dz, X of the detection abject
`detected by the cbiect recognition part to the mobile body on the basis of parallax
`information acauired fram 4 pair of pickup images picked by a stereocamera, and
`apeciies relative speeds Vx, Vy, Vz of the detection oblect to the mobile body from a
`difference between ihe image positions of the detection object between two or more
`pickup images (frames) whose pickup periods are different, An object recognition part
`208 for control determines whether or not the specified relative positions and relative
`speeds satisfy the conditions of the relative positions and relative speeds as an cbiect
`for cortral,
`
`https://worldwide.espacenet.com/publicationDetails/biblio?J=1&ND=3 &adjacent=true&loc... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 1 of 13
`
`Powered by EPO and Google § s
`~~«te,
`
`rg
`
`Patent Transiate
`
`nerated. I cannot be guaranieed thai itis intelligible, accurate, complete, retiable or ft for specific purposes, Critical decisions,
`This translation is machine-ge!
`such as commercially relevant or financial decisions, should not be based on machine-transiation oufsut,
`
`
`
`
`
`WwW Le GaAvg
`GRE TOTISR FS
`
`An oblect of the present invention is ia appropriately identify whether or qot an cbhiecl prasent around a moving object is a
`contral object, and ta enable apprapriate cantrol of a predetermined device mounted on tne moving oblect. An oblect recognition
`unt (205) detects 3 detection target such as another vehicle ar a pedestrian present around a moving body Gasecd on a captured
`
`
`image obtained by Imaging the perghery of the moving body. On the ather hand,
`the oblect slate determination unit 207
`specifies the lative position Dz, X of the detection target detected by the object recognition unt with respect to the moving
`
`obiect based on parallax information obtained from a pair of captured images captured by a stereo camera. The relaiive
`
`velocities Vx, Vy, Vz of the detection langerelative to the moving objact are specified frorn the differance in ihe image pasiien
`of the detection target hetween Iwo or more cantured images
`(between frames} having diferent imaging timings. The contra!
`
`object recognition unit 238 determines whether or not the specified relative position and relative velocity satisfy the conditions of
`ihe rehative pasitien and the relative velocity fo be the contral oblect. [Selected figure} Figure 4
`
`Contral object recognition device, mobile device contral system and cuntrol object recognition program
`
`fond)
`
`The present invention is provided with a control target identification device for identiying whether or not a detection target
`present around a mobile is a contral target for controlling a predetermined device mounted on the mobile, The present inventien
`relates to a mobile device control system and a contro! abject recognition program.
`
`(8002)
`
`Conventionally, as & device for detecting an oblect exisiing around a moving bady, for exam@le, a driver suppart system (mobile
`bady such as ACC (Adaptive Cruise Control for reducing a driving load of a driver of a vehicle (mobile body} What is used for
`
`ment contral systems etc. is Known.
`
`
`The driver assistanoe system cantrois varous devices mauntacd on the host vehicle, andi has an autornatic brake funcilon and an
`alarm for preventing the host vehicle from colliding with an obstacie or the like and reducing the impact in the event of a
`collision.
`it realizes various functions such as function, sel-vehicie speed adjustment function to maintain inter-vehicie distance
`with leading vehicis, and functian fo sucpart prevention of depariure fram ihe tane where both vehicies are traveling.
`
`i003)
`
`in Patent Document 7, distance data (parallax value) is calculated from a pair of captured images obtained by capturing an area
`in frant of the hast vehicie by a sierea carmera mounted an the hast vehicle, and above the road surface based on the distance
`data Discluses = three-dimensional object detection apparatus for detecting a three-dimensional object present in in this three-
`dimensional oblect detection device, the three-dimensional object detected in the processing target Image (one of the pair of
`captured images} captured af a certain time is regarded as a stationary obiect, and the motion state of the steres camera Using
`ui
`
`the detection result
`of the vehicle speed information and the distance data of the three-dimensional obiect,
`the image
`
`coordinates at which the three-dimensional oblect aonears at the next dme are oredicted. When the three-dimensional abject is
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 2 of 13
`
`a stationary object, the diference between the luminance value of the image area of the three-dimensional obiect and the
`luminance value of the image coordinates related io ihe grediction of the image actually captured at the next time is zero. When
`the three-dimensional oblect is a moving abiect, the difference is not zero. Thus, according fo Patent Document 1, 1 is passibie
`to detect a moving obiect moving in the front region of the host vehicle at a high processing speed.
`
`i004}
`
`is necessary fo contra!
`ft
`in order to properly function a mobile device control sysiem such as a driver assistance system,
`predetermined devices on the mobile obiect controlled by the mobile device contre] system by various objects shown in the
`captured image. itis imporiant to accurately identify whether it is a control object or not. Han obiect that is a non-contral ohiect
`is recognized as a control object,
`if causes a malfunction that performs device contra! thal is not necessary originally, and
`conversely, an obiect that is a contro! object is not if itis recognized that Kis an object for control, a necessary device contra!
`may not be axeculed.
`
`Ogos}
`
`However, depending on the control content of the device on the moving object, even the same object may or may not be the
`
`
`contral object depending on the specific behavior of ihe oblect to the moving obleci. Far exampie, when the control content is an
`
`aulomatic braking function er an alarm function, when the preceding vehicie shows a oehavior traveling at a speed equal to or
`higher than that of the own vehicle, the preceding vehicle obstructs the traveling of the own vehicle Or since if
`is not an
`avoidance thing fo aveid, ii does noel become a cantrol object. On the other hand, when the oreceading vehicle traveling in frant of
`ihe awn vehicle shows a behavior of rapid deceleration ar shows a behavior of rapidly intruding into the front af the own vehicle,
`the preceding vehicis is an obstacie or obstacie that prevenis the traveling ofthe own vehicle it becomes an avoidance obiect to
`be avalned, and becornes an object fo be cantrolled such as the automatic brake funcion and the alarm function. Therefore, in
`order fo perform various contrals, iis desirable io be able to comprehend the specific behavioral content of the mabile abject.
`Although the three-dimensional chject detection devine disciosed in Patent Document 7 has described the detection of a moving
`ebiect shawn ina captured image, if does not touch on grasping the seecific Gehavier content of the moving oblect wHh respect
`to the moving object.
`
`roooe)
`
`Thea greseqdi invention has been made in view of the above background, and an Gbiect of the orasent invention is io oroperly
`identify whether or not an object existing around a mubile object is a control target. and fo mount the mobile object on the mobile
`object. His an chject of ihe oresent invention to provide a contra: object recognition device, a mobile device control system and
`8 codiral abject recognition program thal enable appropriate control af devices,
`
`roa?)
`
`in order to achieve the above obiect, the present invention comprises an object detection means for detecting a detection object
`present around ihe moving obiec! based an a captured image oGbainad Oy imaging the surroundings of the moving object, and
`the object A contro! target identification devices that identifies whether a detection target detected by a detection uniiis a contra!
`target that controls a predetermined device mounted on the mobile object, and iniages ihe periphery of the mobile object. The
`rigive oosition spectying means for specifying the relative pasion of the detection ablact detected by the object detection
`means with respect te the movable body based on the paralax information cbtained from the pair of captured images imaged by
`
`
`the giuralty of imaging means, and the imaging me Relative velocity specifying means for specifying the relative velocity of the
`
`detection ablect ia the moving object from ihe difference in the image position of ihe delection otfect between two or mare
`different captured images, and the relative postion identification means ft characterized by having a determining means which
`réiative speed is location and the relative velocity specification unit identified to determine whether or not the condition of the
`rigiive pasitian and the relive scead to be the coniral subject.
`
`(00s)
`
`if is possible to soppropriately identty whether ar nat an abject present around the moving
`According to the present invention,
`
`oblect is a conirel target, anc te achieve an excelled: effect of enabling appropriate control of a predetermined device mounted
`on the moving object. Fu.
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 3 of 13
`
`ooog]
`
`His a schematic diagram which shows schematic siructure ofa vahicle-mounted apparatus cantral system.
`
`itis a schematic diagram which shows schematic structure of the Imaging unit which comerises the vehicie-mounted apparatus
`control system, and an imaqge-analysis unit,
`
`itis a mode! eniarged view when the image sensor in the imaging part of the imaging unit is seen fram the direction orthogonal!
`to the Hight transmission direction.
`
`
`His a block dlagrarn showing compasiicn of a contral subject identification device in an embadiment. His a flow chan which
`
`shows 2 flow of identification processing of a control subject. (A) And (b) is a figure which shows the lefi-right image of a sterea
`camera, respectively. i is a figure exgiaining the principle of ranging calculation. tis an explanatory view showing an otect
`candidate Held set up in an image pick-up picture duminance picture). ft
`is explanatory drawing which shows the identificatian
`method of the oblect candidate area | region. (A) is the figure which showed typically the parallax image corresponding to the
`captured image shown in FG. (B) is a figure which showed typically the horzentai line parallax histogram Gf mam) obtained
`frora the parallax image of ine figure fa). (2) is explanatory drawing which Hlustrated the read surface area | region in ine
`parallax image of the figure fa).
`Ht
`is explanatory drawing which shows four types of feature pattern examples used for
`identification of a detection target object. H is explanatory drawing which shows the structure of the identifier used by the
`recognition process of a detection target object. (A) is explanatory drawing which shaws image area A, ©, E of each detection
`target chiect recognized by this object recognition process. (B} is explanatory drawing which shows tracking resull area | region
`A’, Gand § lof each detection target object in the following flame | frame. (A) is a figure which shows iygically the parallax
`image corresponding te the captured image duminance image} shown fo Fig.13 (4). (8) is a figure which shows typically the
`paraliax image corresponding to the captured image duminance image) shown in FRS.13 (bp). FIG. 6 is an explanotory diagram
`for descdbing siate feature quaniiies (XA, D2 A, VxA, Vy A, ¥ 2 A) ofa detection target Goreceading vehicie} displayed in an
`image area A.
`
`foo te}
`
`Hereinafter, an embodiment in which a cantrol object klentificaiion device according fo ihe present invention is used in an on-
`vehicie device cantrol system as a mobile device control sysiem will be described. FRG.
`1
`is a schematic view showing 2
`schematic configuration of an in-vehicie device control system. The in-vehicle device control system shown in FIGS. 4
`is provided
`wih an imaging uni 1OQ4 as an imaging unit thal images a region in fant of the traveling vehicle 190 In ihe traveling direction as
`
`an imaging region. The imaging unit 704 is installed, for example,
`in the vicinity of a rearview mirror (not shown) of the
`windshield 103 of the vehicle 146. Various dota such as captured image date obtained by imaging of the imaging unit 101 is
`
`
`input ig an image analysis un —_ 102 as an image processing uqdk. The inage analysis uqi 02 analyzes the data transmitted
`fram the iraaging un fi 101 to calculate the position, direction, and distance of the other vehicie present ahead of the host vehicie
`100, of on ihe road surface existing in the imaging region. Detect lane boundaries such as white lines.
`In the detection of
`another vehicle, an object on a road surface is deiected as 2 veticle based on a parallax image.
`
`ort
`
`The calculation result of the image analysis uni 72 is also sent io the vehicle travel control unt 1o4. The vehicie travel contral
`unt 104 drives ihe subject vehicle 700, far example, when ihe subject vehicle 700 is likely fo collide wih an abstacie, based on
`ihe detection results af the defection target such as a pedestrian or a traveling vehicle detected by the image analysis unit 102.
`informing @ person of a warning or performing driving Support control such as controtting the steering wheel and the brake of the
`wn verichs.
`
`{0042}
`
`FES. 2 is a schematic view showing a schematic configuration of the imaging unit 104 and the image analysis uni 702. As
`shown iq PYG. The imaging unl (O17 is @ stereo camera including hwo imaging unis 1104 and 1705, and the canfigurations of
`
`the two imaging unfs 1704 and 4108 are the same. The imaging unis 71OA and 1168 respectively include imaging lenses
`TTA and 1478, image sensors 1724 and 1128 In which imaging elements are hwo-dimensionally arranged, sensor substrates
`
`TIiSA and 7138 provided with the image sensors 1724 and T12B, and a sensor substrate Signal processing units 1744 and
`1145 that generate and outoul captured image data obtained by converting analog electrical signals (light reception amounts
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 4 of 13
`
`received by light receiving elements on the image sensors 1424 and 1128) output from 1134 and 1138 into dighal electrica!
`signals it consists of Luminance Image data is output fromthe imaging unit 194.
`
`10043]
`
`The imaging unii 107 aise includes = processing hardware unit 120 configured of an FPGA (Field-Programmabie Gate Array} or
`
`the like.
`in order to obtain parallax images from the iurninance image data output from the imaging unts TIGA and 1108, the
`processing hardware unit 120 calculates the caratlax values of corresponding image portions between the imaged images
`captured by the imaging units 170A and 110B. A disparity calculation unii t21 that calculates is orovided. The parallax value
`referred to here is a comparison iniage with respect to an image portion on the reference image corresponding to the same
`point in the imaging region, with one af the iriages taken by each of the imaging units 110A and 1108 as the reference image
`and the other as the comparison image. The positional displacement amount of the upper image portion is calculated as the
`
`
`parallax value of ihe image portion. By using the principle of tdangquiatien,
`if
`is passigie to calculate the distance from the
`parallax value ta the sare point in the imaging regien correspanding to the image portion.
`
`10044]
`
`Onihe other hand, the image anahysis unt 702 includes a memory 139 for sloring furinance image data and paralkix image
`
`data cuiput fram the ‘tnaaing uni ¢O7, and an MPU (Microcomputer that incorserates sofiware for perfarming identificatian
`
`nrocessing of control target oblects and parallax calculation control, Processing Uni 140. The MPU 140 executes identification
`
`
`sessing ofacontrol targ bject nding to the present embodiment g ihe luminance ima Jala and the paralla
`
`
`
`
`
`
`3ocessing
`of a cal
`arget o
`according ta the
`present embodiment usin
`ening
`wge
`data
`and
`the parallax
`image data stored in the memory £20.
`
`foots}
`
`FIG, 3 is a schernatic aniarged view of the image sensors 772A and 14123 when viewed from the direction arihagonal fo the light
`transmission direction. Fhe image sensors 112A and 112B are image sensors using a charge coupled device (OCD), a
`
`complementary metal oxide semiconductor (OMOS}, or the ke, and a photodiode tiga is used as an imaging element dight
`receiving element. The pholodiodes 112 4 are iwo-dimensionally anayed for each imaging pixel, and in order to increase the
`light collectian efficiency ofthe photodiodes 412 a, a micro fens 172 0 is provided on the incident side ofeach ghutodiode 112 a
`The image sensors 1124 and 112B are bonded io 4 printed wiring Goard (PYVB} by a method such as wire bonding fo form
`sensor subsirates 7234 and $132.
`
`Ogts}
`
`
`Next, cantrol object identification processing, which is a feature of the present invention, wil be deseribed. FRG. 4 is a block
`diagram showing ihe configuration of the contral dargel xieqtiicaiion apparaius of the present ernbadiment realized by t
`processing hardware unit 120 and the image analysis unl 102 in FIG. In the figure, the control object identification apparatus
`290 of the present embodiment includes a stereo image input unit 204, a paraliax Image calculation unit 202, a lurunance image
`igpul unt 203, an oblect candidate area recognition uri 204, an obiect racognitien unit 205, and an object iracking process. His
`configured to include a unit 208, an object state determination unit 207, a contro! target recognition unt 208, a recognition
`dichonary storage unt 209, and a recognition result output unit 270.
`
`10087}
`
`A stereo image is input fo the stereo image input unit 204 from a stereo camera (not shown) provided with an imaging unit
`having an imaging lens and an image sensor on the left and right. A luminance image which Is a left image ora right image of a
`stereo image is input te the luminance image inoul unlf 203 from ihe sharac image Inout unit 204, The input sterec image and
`luminance image are stered in the memory 130 of the sterea camera. The parallax image calculation unit 202 calculates a
`parallax value, which is 3 difference between the imaging positions af the left and right Images, using the stereo image input to
`
`ihe sterea image Input uni 204. and sets the parallax value as a pixel value. Caiculaie ihe made.
`
`Ootg}
`
`The abject candidate area recognition unit 204 detects a road surface area from the captured image using the parallax Image
`calculated by the parallax image caloulatien unt 202, and the defection reasudi of the road surface area and the luminance image
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 5 of 13
`
`input from the luminance image input unit 203 To set an ohiect candidate area incliding the detection target. The object
`recognition unit 205 perferms recognition processing of the detection target on the obiect candidate area set by the obect
`candidate area recognition unt 204. The object tracking processing unt 208 performs tracking processing on tne detection
`
`target abject
`
`rec:vegized by the obiect
`
`determination unif 2O7 uses the resull of tracking by the ablect tracking processing unit 206 te obtain a motion feature amount
`such as the relative cosition or relative velocity of the detection target.
`
`recognition unit 208 using the next
`
`luminance image frame. The object state
`
`ots}
`
`The recognition dictionary storage unit 200 is a mechine leaming methad such as SVM (Suppor Vector Machine}, and is
`dictionary data for recogniticn of a control target created using image sample learning data of the control target in advance
`
`{recognition dictionary Remember). The recagnitisn dictlanary storage unf 209 stores recagnition dictionaries ee
`
`created for each control ablect. The control target recognition unit 208 pedarrns cantrol target resagnition processing wit
`reference fo the recognition dictionary stored in the recognitien dictlonary storage unl 209 using the behavior feature amount
`obtained by the cbiect state determination unl 207... The recognition result output unit 270 cutouts the recognition result of the
`cantrol Gbject racognitian unt 208.
`
`og2a}
`
`Next, the operation of the object recognition device according io the present embadiment will bs described according to FIG. 5
`showing the operation flaw. A sterea image is Inout ia the stereo image Inout unk 201 of FRG. 4 (3404). Specifically, a sterea
`image is Inout fram a stereo camera. FIG. 6 shows an example of a stereo image. In stereo images, the same subject is imaged
`ai different imaging positions by the left and right image sensors. Then, the luminance image Input unt 209 of Fic. 4 inputs a
`luminance image (refarance image) of eliher the lef image or ihe right image of ine stereo image (S192). The input stereo
`image and luminance image are stored in the memory 130 of the stereo camera.
`
`os]
`
`The oblect candulate area recognition unt 204 seis an object candidate area ina place where an cblect exisis based on the
`luminance image (2103). Specifically, first,
`in order to set a candidate region of a detection target, as shown in FIG. & a
`reclangular Glock is set according io the image of the detection iarget in the captured image. The sosition of the rectangular
`
`blockin the captured image and ihe she of ihe rectangular block are specified by the upperleft coordinates Ofs, Ys} and lower
`
`fight coardinates (Xe, Ye) of the rectangle as shown in Fic. When seiting rectangular blocks, the rectangular blacks are set in
`
`
`order fromthe largest rectangular block. This rectangular block is an Gbiect candidate area.
`
`10022]
`
`The parallax image calculation unit 202 calculates the carallax value which is the difference between the imaging positions of
`the left and right images in ihe target frame using the stereo image input te the stereo image input unit 207 (6104), Specifically,
`for ihe same cortion af the sterac image formed oy the lef and right imading lanses, the parallax value is calculated by ihe block
`matching method, and the parallax image with the calculated parallax value as the olxel value is calculated. The block matching
`method is a method of dividing left and night images into blocks, and when the bisok similarity in the left and right images is the
`
`largest, oolaining a paralax value from a portien where the blocks are muiched. For exarnpis, a 1280 « 960 pixel image is
`divided info 5 x & sized blocks. The optimal value of block size is adjusted and set by experiment. in the examole shown in FRG.
`
`7, the imaging positions tn the lef and night images with respect fo the paint © on the subject are the distances Ad and Az from
`
`the imaging center, Therefore, the parallax value A is calculated as A = Ad + Ae. Therefore, the parallax image is an image
`having parallax values at pixel positions.
`
`(0023)
`
`
`The ogiect candidate area racognitien unit 204 creates a horizontal line parallax histograrn (¥ mag) showniq FIG. 105 using the
`parallax image obtained by the parallax image calculation unit 202 (S405). The horzontal line parallax histogram js a histugram
`
`obtained by integrating the parallax values included in ihe parallax Image in one horizontal line of the parailiax image in FRG.
`
`That is, FG. 10 (6) shows the frequency value distributiog H CA, y) al the paraiiax value A. The horizontal axis of FG. TOG fs the
`parallax value A, and the vertical axis is the height y of the parallax image. The value of each pixel of the horizontal line parallax
`
`histogram is the frequency of the parallax value of oneling of the height y af the parallax image.
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 6 of 13
`
`foo24)
`
`Tha road surface area projeciad oy ihe cagturad image tends to decrease in width at a suostantially constant rate from the lower
`par to the upper cart of the image. Then, the road surface partion shawn on the same horizontal line on the captured image has
`substantially the same parallax value A on the parallax image because the distance fram the host vehicle 190 is substantially
`
`ihe same, Therefore, the gixels corresponding to the road surface area have a Hnear distrigution having an inclinaiion as shown
`in FIG, 198 on the horizontal line garallax histogram.
`
`{0025}
`
`line parallax
`Therefore, ihe otfect candidaie area recognition unit 204 firsi detects pixels linearly arranged on the horkmonial
`histogram, and mags the detected pixels on the parallax image. Thal is, the pixels Hnearly arranged in the horizontal
`ti
`
`paraiiax histogram are associated as pixels of the road surface area on the parallax image. Thus,
`if the pixel which projects a
`mad sudiace is known, a road sundiace area 302 as shown in FRG.19 ©) can be detested by perfarming an interoolation process
`between the pixels (8106).
`
`26}
`
`Specifically, the ablect candidate araa recogrilian unit 204 accuiras: paraiiax distributien infarmation in aach harizontal directian
`fram ihe parallax image calculation unit 202, and the pixel distribution on the horizontal line parallax histegram specified fram
`the information is obtained by the least squares method of the existing technology or Linear approximation is performed by the
`Haugh transform method. An approximate siraight line as shown in FIG. 108 ablained by this becomes a straight line having
`such an inclination that the parallax value becomes smatiier toward the upper side of the image in the horizontal line parallax
`histogram corresponding to the lower part ofthe parallax image. The oixels distributed on the approximate straight line or in the
`vicinHy thereof are present al substantially the same distance in each hodzontal line on ihe parallax image and have a high
`occupancy ratic, and a target of which the distance is continuausly farther away foward the unper side of the image Ht is a pixel.
`That is, ihe pixeis distibuted on the approximate straight line in the horizontal line parallax histogram or in the vicinity thereof
`incicate thai the roaci surface area is projected, As a result, ihe mac suriace area S02 can be detected by specifying the pixels
`
`distributed on the approximate straight line in the horizontal Hne parallax histogram or in the vicinity thereof, Although the road
`surface aréa is detected by the above method, the road shoulder and ihe white line are recognized framthe luminance image,
`and the road surface ares is detected by specifying the continuous area of the road shoulder and the area where the white Hne
`exists. You may
`
`{0627}
`
`Here, inthe present embodiment, an opstacie that hinders the traveling of the hast vehicie 100, an object fo be avaided, and the
`like are detected, and control is performed ic notify the driver of the host vehicle 100 of a warning. And driving support contra!
`
`
`such as controling ihe brakes. An oblect (contral oblect) to be detected fo realize these controlsis an object present on ihe road
`surface on which the vehicle 190 is traveling.
`
`{0028}
`
`Therefore, in the present embodiment, when ihe road surface area 302 is detected, the obiect candidate area recognition unit
`
`204 sets the recognition weight of ine road area ta 1 and the recognition weight of the other image areas fo 0 (S707). Then, an
`oblect candidate area overlapping the area whose recognition weight is 1
`is an oblect candidate area which is described later on
`the road surface as an object candklate area existing on the road surface, while an ohiect not ovedapping the area whose
`recognition weight is ¢ The oblect recognifion process described later is nat pedurmed on the candidate area as the candidate
`area of the object not existing on the road surface (S108). Therefore,
`in the example shawn in FES. 8c, object recognition
`precessing is gerformed on object candidaie regians indicaled by reference numerais A, C, and ©, and abject recagnition
`processing is performed on object candidate regians indicated by reference numerais B,D, and F.
`iH dues not happen. As a
`result, in addition to the possibility of reducing false recogniticn In which an obiect that is not the contra! farget is errongously
`recognized as the cantrol hargei, the qurnberof object candidate areas for which the oblect recognition processing is pedormed
`is reduced. .
`
`ooegi
`
`http://translationportal.epo.org/emtp/translate/?ACTION=description-retrieval& COUNTRY... 6/9/2019
`
`

`

`2015133078-Desc-en
`
`Page 7 of 13
`
`The oblect recognition uni 205 perferms an ohiect recognition process of the detection target on an obiect candidate region
`overlapping ihe region having a recognition weight of ¢ among the object candidate regions set by the cbiect candidate region
`recognition unit 204 ($109).
`in this oblect recognition process, for example, a recegniiion dictlonary is created using image
`sampie learning data on which an object is recognized in advance by a machine learning method. Af this time, it is necessary to
`create recagnition dictionaries separately depending on the detection langet.
`
`(0030)
`
`First, the obiect recognition unit 205 sets the white region 404 and the black region 402 In the four types of black and white
`
`blocks Ato D as shown in FIG. Tf for the object candidate region set by the object candidate region recognition unit 204 (see
`FES. Calculate the feature quaniities of Undicated by middle hatching). Specifically, the surn of ihe pixe! values for the oorion in
`the obiect candidate area corresponding to the white area 401 of cach black and white block 400 of Ato 5 as shown in FRG. 11
`and the abject candioaie area carresconding to the black area 402 Calculate the sum of pixe! values for the lacation of. Then, a
`value obtained by subtracting the oixel value sumin the black area 402 from the gixe! value sum in the white area 407 is set as
`
`the feature amount h Og regarding ihe type of black and white block for the object candidate area.
`
`10034]
`
`One evaluation value f Gd is calculated from the feature amount h Og calculated in this manner, using the evaluation functian
`shown in the following equation (1). The evaluation value f Od is a feature amount ht Gg} for cach black and white block @ = 4 to
`¥ (Poo 4 in fhe case of using four black and while blocks Hhuistraied in FIG. f¢}}. To calculate by surmrning the weighted feature
`quantities a t= ht Og of the respective black and white blacks obtained by mullipiying the weight calculation results by the
`weighting factor at given for each black and white block. te can. Here, the feature amount ht OO and the weighting factor o t can
`be obtained by collecting learning data for the luminance image of the detection target and leaming fram the learning dala.
`
`£00223
`
`
`As shown in FiG. 12, the classiier used in the recognition process of such a detection target is configured to include a plura
`of levers 500-4 ion 1 is 4 positive inleged. Each of the layers 500-7 io 500-n has individual evaluation functions using differant
`
`feature sets features h t &) and weighting factors at) that are characterized as being specific detection abiects. ing. Then, for
`the object candidate area, first, an evaluation va

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket