Committer: Michael Beasley <mike@snafu.setup>
[mikesnafu-overlay.git] / drivers / infiniband / hw / ehca / ehca_av.c
blob194c1c30cf63c4d2fd16a75619e6d4ec2aac4dc1
1 /*
2 * IBM eServer eHCA Infiniband device driver for Linux on POWER
4 * address vector functions
6 * Authors: Hoang-Nam Nguyen <hnguyen@de.ibm.com>
7 * Khadija Souissi <souissik@de.ibm.com>
8 * Reinhard Ernst <rernst@de.ibm.com>
9 * Christoph Raisch <raisch@de.ibm.com>
11 * Copyright (c) 2005 IBM Corporation
13 * All rights reserved.
15 * This source code is distributed under a dual license of GPL v2.0 and OpenIB
16 * BSD.
18 * OpenIB BSD License
20 * Redistribution and use in source and binary forms, with or without
21 * modification, are permitted provided that the following conditions are met:
23 * Redistributions of source code must retain the above copyright notice, this
24 * list of conditions and the following disclaimer.
26 * Redistributions in binary form must reproduce the above copyright notice,
27 * this list of conditions and the following disclaimer in the documentation
28 * and/or other materials
29 * provided with the distribution.
31 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
32 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
33 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
34 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
35 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
36 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
37 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
38 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
39 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
40 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
41 * POSSIBILITY OF SUCH DAMAGE.
45 #include <asm/current.h>
47 #include "ehca_tools.h"
48 #include "ehca_iverbs.h"
49 #include "hcp_if.h"
51 static struct kmem_cache *av_cache;
53 int ehca_calc_ipd(struct ehca_shca *shca, int port,
54 enum ib_rate path_rate, u32 *ipd)
56 int path = ib_rate_to_mult(path_rate);
57 int link, ret;
58 struct ib_port_attr pa;
60 if (path_rate == IB_RATE_PORT_CURRENT) {
61 *ipd = 0;
62 return 0;
65 if (unlikely(path < 0)) {
66 ehca_err(&shca->ib_device, "Invalid static rate! path_rate=%x",
67 path_rate);
68 return -EINVAL;
71 ret = ehca_query_port(&shca->ib_device, port, &pa);
72 if (unlikely(ret < 0)) {
73 ehca_err(&shca->ib_device, "Failed to query port ret=%i", ret);
74 return ret;
77 link = ib_width_enum_to_int(pa.active_width) * pa.active_speed;
79 if (path >= link)
80 /* no need to throttle if path faster than link */
81 *ipd = 0;
82 else
83 /* IPD = round((link / path) - 1) */
84 *ipd = ((link + (path >> 1)) / path) - 1;
86 return 0;
89 struct ib_ah *ehca_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
91 int ret;
92 struct ehca_av *av;
93 struct ehca_shca *shca = container_of(pd->device, struct ehca_shca,
94 ib_device);
96 av = kmem_cache_alloc(av_cache, GFP_KERNEL);
97 if (!av) {
98 ehca_err(pd->device, "Out of memory pd=%p ah_attr=%p",
99 pd, ah_attr);
100 return ERR_PTR(-ENOMEM);
103 av->av.sl = ah_attr->sl;
104 av->av.dlid = ah_attr->dlid;
105 av->av.slid_path_bits = ah_attr->src_path_bits;
107 if (ehca_static_rate < 0) {
108 u32 ipd;
109 if (ehca_calc_ipd(shca, ah_attr->port_num,
110 ah_attr->static_rate, &ipd)) {
111 ret = -EINVAL;
112 goto create_ah_exit1;
114 av->av.ipd = ipd;
115 } else
116 av->av.ipd = ehca_static_rate;
118 av->av.lnh = ah_attr->ah_flags;
119 av->av.grh.word_0 = EHCA_BMASK_SET(GRH_IPVERSION_MASK, 6);
120 av->av.grh.word_0 |= EHCA_BMASK_SET(GRH_TCLASS_MASK,
121 ah_attr->grh.traffic_class);
122 av->av.grh.word_0 |= EHCA_BMASK_SET(GRH_FLOWLABEL_MASK,
123 ah_attr->grh.flow_label);
124 av->av.grh.word_0 |= EHCA_BMASK_SET(GRH_HOPLIMIT_MASK,
125 ah_attr->grh.hop_limit);
126 av->av.grh.word_0 |= EHCA_BMASK_SET(GRH_NEXTHEADER_MASK, 0x1B);
127 /* set sgid in grh.word_1 */
128 if (ah_attr->ah_flags & IB_AH_GRH) {
129 int rc;
130 struct ib_port_attr port_attr;
131 union ib_gid gid;
132 memset(&port_attr, 0, sizeof(port_attr));
133 rc = ehca_query_port(pd->device, ah_attr->port_num,
134 &port_attr);
135 if (rc) { /* invalid port number */
136 ret = -EINVAL;
137 ehca_err(pd->device, "Invalid port number "
138 "ehca_query_port() returned %x "
139 "pd=%p ah_attr=%p", rc, pd, ah_attr);
140 goto create_ah_exit1;
142 memset(&gid, 0, sizeof(gid));
143 rc = ehca_query_gid(pd->device,
144 ah_attr->port_num,
145 ah_attr->grh.sgid_index, &gid);
146 if (rc) {
147 ret = -EINVAL;
148 ehca_err(pd->device, "Failed to retrieve sgid "
149 "ehca_query_gid() returned %x "
150 "pd=%p ah_attr=%p", rc, pd, ah_attr);
151 goto create_ah_exit1;
153 memcpy(&av->av.grh.word_1, &gid, sizeof(gid));
155 av->av.pmtu = shca->max_mtu;
157 /* dgid comes in grh.word_3 */
158 memcpy(&av->av.grh.word_3, &ah_attr->grh.dgid,
159 sizeof(ah_attr->grh.dgid));
161 return &av->ib_ah;
163 create_ah_exit1:
164 kmem_cache_free(av_cache, av);
166 return ERR_PTR(ret);
169 int ehca_modify_ah(struct ib_ah *ah, struct ib_ah_attr *ah_attr)
171 struct ehca_av *av;
172 struct ehca_ud_av new_ehca_av;
173 struct ehca_pd *my_pd = container_of(ah->pd, struct ehca_pd, ib_pd);
174 struct ehca_shca *shca = container_of(ah->pd->device, struct ehca_shca,
175 ib_device);
176 u32 cur_pid = current->tgid;
178 if (my_pd->ib_pd.uobject && my_pd->ib_pd.uobject->context &&
179 my_pd->ownpid != cur_pid) {
180 ehca_err(ah->device, "Invalid caller pid=%x ownpid=%x",
181 cur_pid, my_pd->ownpid);
182 return -EINVAL;
185 memset(&new_ehca_av, 0, sizeof(new_ehca_av));
186 new_ehca_av.sl = ah_attr->sl;
187 new_ehca_av.dlid = ah_attr->dlid;
188 new_ehca_av.slid_path_bits = ah_attr->src_path_bits;
189 new_ehca_av.ipd = ah_attr->static_rate;
190 new_ehca_av.lnh = EHCA_BMASK_SET(GRH_FLAG_MASK,
191 (ah_attr->ah_flags & IB_AH_GRH) > 0);
192 new_ehca_av.grh.word_0 = EHCA_BMASK_SET(GRH_TCLASS_MASK,
193 ah_attr->grh.traffic_class);
194 new_ehca_av.grh.word_0 |= EHCA_BMASK_SET(GRH_FLOWLABEL_MASK,
195 ah_attr->grh.flow_label);
196 new_ehca_av.grh.word_0 |= EHCA_BMASK_SET(GRH_HOPLIMIT_MASK,
197 ah_attr->grh.hop_limit);
198 new_ehca_av.grh.word_0 |= EHCA_BMASK_SET(GRH_NEXTHEADER_MASK, 0x1b);
200 /* set sgid in grh.word_1 */
201 if (ah_attr->ah_flags & IB_AH_GRH) {
202 int rc;
203 struct ib_port_attr port_attr;
204 union ib_gid gid;
205 memset(&port_attr, 0, sizeof(port_attr));
206 rc = ehca_query_port(ah->device, ah_attr->port_num,
207 &port_attr);
208 if (rc) { /* invalid port number */
209 ehca_err(ah->device, "Invalid port number "
210 "ehca_query_port() returned %x "
211 "ah=%p ah_attr=%p port_num=%x",
212 rc, ah, ah_attr, ah_attr->port_num);
213 return -EINVAL;
215 memset(&gid, 0, sizeof(gid));
216 rc = ehca_query_gid(ah->device,
217 ah_attr->port_num,
218 ah_attr->grh.sgid_index, &gid);
219 if (rc) {
220 ehca_err(ah->device, "Failed to retrieve sgid "
221 "ehca_query_gid() returned %x "
222 "ah=%p ah_attr=%p port_num=%x "
223 "sgid_index=%x",
224 rc, ah, ah_attr, ah_attr->port_num,
225 ah_attr->grh.sgid_index);
226 return -EINVAL;
228 memcpy(&new_ehca_av.grh.word_1, &gid, sizeof(gid));
231 new_ehca_av.pmtu = shca->max_mtu;
233 memcpy(&new_ehca_av.grh.word_3, &ah_attr->grh.dgid,
234 sizeof(ah_attr->grh.dgid));
236 av = container_of(ah, struct ehca_av, ib_ah);
237 av->av = new_ehca_av;
239 return 0;
242 int ehca_query_ah(struct ib_ah *ah, struct ib_ah_attr *ah_attr)
244 struct ehca_av *av = container_of(ah, struct ehca_av, ib_ah);
245 struct ehca_pd *my_pd = container_of(ah->pd, struct ehca_pd, ib_pd);
246 u32 cur_pid = current->tgid;
248 if (my_pd->ib_pd.uobject && my_pd->ib_pd.uobject->context &&
249 my_pd->ownpid != cur_pid) {
250 ehca_err(ah->device, "Invalid caller pid=%x ownpid=%x",
251 cur_pid, my_pd->ownpid);
252 return -EINVAL;
255 memcpy(&ah_attr->grh.dgid, &av->av.grh.word_3,
256 sizeof(ah_attr->grh.dgid));
257 ah_attr->sl = av->av.sl;
259 ah_attr->dlid = av->av.dlid;
261 ah_attr->src_path_bits = av->av.slid_path_bits;
262 ah_attr->static_rate = av->av.ipd;
263 ah_attr->ah_flags = EHCA_BMASK_GET(GRH_FLAG_MASK, av->av.lnh);
264 ah_attr->grh.traffic_class = EHCA_BMASK_GET(GRH_TCLASS_MASK,
265 av->av.grh.word_0);
266 ah_attr->grh.hop_limit = EHCA_BMASK_GET(GRH_HOPLIMIT_MASK,
267 av->av.grh.word_0);
268 ah_attr->grh.flow_label = EHCA_BMASK_GET(GRH_FLOWLABEL_MASK,
269 av->av.grh.word_0);
271 return 0;
274 int ehca_destroy_ah(struct ib_ah *ah)
276 struct ehca_pd *my_pd = container_of(ah->pd, struct ehca_pd, ib_pd);
277 u32 cur_pid = current->tgid;
279 if (my_pd->ib_pd.uobject && my_pd->ib_pd.uobject->context &&
280 my_pd->ownpid != cur_pid) {
281 ehca_err(ah->device, "Invalid caller pid=%x ownpid=%x",
282 cur_pid, my_pd->ownpid);
283 return -EINVAL;
286 kmem_cache_free(av_cache, container_of(ah, struct ehca_av, ib_ah));
288 return 0;
291 int ehca_init_av_cache(void)
293 av_cache = kmem_cache_create("ehca_cache_av",
294 sizeof(struct ehca_av), 0,
295 SLAB_HWCACHE_ALIGN,
296 NULL);
297 if (!av_cache)
298 return -ENOMEM;
299 return 0;
302 void ehca_cleanup_av_cache(void)
304 if (av_cache)
305 kmem_cache_destroy(av_cache);